Search Results for author: Sebastian Vollmer

Found 14 papers, 5 papers with code

Quantitative knowledge retrieval from large language models

1 code implementation12 Feb 2024 David Selby, Kai Spriestersbach, Yuichiro Iwashita, Dennis Bappert, Archana Warrier, Sumantrak Mukherjee, Muhammad Nabeel Asim, Koichi Kise, Sebastian Vollmer

Large language models (LLMs) have been extensively studied for their abilities to generate convincing natural language sequences, however their utility for quantitative information retrieval is less well understood.

Imputation Information Retrieval +2

X Hacking: The Threat of Misguided AutoML

no code implementations16 Jan 2024 Rahul Sharma, Sergey Redyuk, Sumantrak Mukherjee, Andrea Sipka, Sebastian Vollmer, David Selby

Explainable AI (XAI) and interpretable machine learning methods help to build trust in model predictions and derived insights, yet also present a perverse incentive for analysts to manipulate XAI metrics to support pre-specified conclusions.

AutoML Interpretable Machine Learning

Energy-Based Models for Functional Data using Path Measure Tilting

1 code implementation4 Feb 2022 Jen Ning Lim, Sebastian Vollmer, Lorenz Wolf, Andrew Duncan

Their ability to incorporate domain-specific choices and constraints into the structure of the model through composition make EBMs an appealing candidate for applications in physics, biology and computer vision and various other fields.

Avoiding C-hacking when evaluating survival distribution predictions with discrimination measures

1 code implementation9 Dec 2021 Raphael Sonabend, Andreas Bender, Sebastian Vollmer

In this paper we consider how to evaluate survival distribution predictions with measures of discrimination.

Survival Analysis

Mitigating Statistical Bias within Differentially Private Synthetic Data

no code implementations24 Aug 2021 Sahra Ghalebikesabi, Harrison Wilde, Jack Jewson, Arnaud Doucet, Sebastian Vollmer, Chris Holmes

Increasing interest in privacy-preserving machine learning has led to new and evolved approaches for generating private synthetic data from undisclosed real data.

Privacy Preserving

Foundations of Bayesian Learning from Synthetic Data

no code implementations16 Nov 2020 Harrison Wilde, Jack Jewson, Sebastian Vollmer, Chris Holmes

There is significant growth and interest in the use of synthetic data as an enabler for machine learning in environments where the release of real data is restricted due to privacy or availability constraints.

Synthetic Data Generation

Debiasing classifiers: is reality at variance with expectation?

no code implementations4 Nov 2020 Ashrya Agrawal, Florian Pfisterer, Bernd Bischl, Francois Buet-Golfouse, Srijan Sood, Jiahao Chen, Sameena Shah, Sebastian Vollmer

We present an empirical study of debiasing methods for classifiers, showing that debiasers often fail in practice to generalize out-of-sample, and can in fact make fairness worse rather than better.

Fairness

Multilevel Monte Carlo for Scalable Bayesian Computations

no code implementations15 Sep 2016 Mike Giles, Tigran Nagapetyan, Lukasz Szpruch, Sebastian Vollmer, Konstantinos Zygalakis

In contrast to MCMC methods, Stochastic Gradient MCMC (SGMCMC) algorithms such as the Stochastic Gradient Langevin Dynamics (SGLD) only require access to a batch of the data set at every step.

Multilevel Monte Carlo methods for the approximation of invariant measures of stochastic differential equations

no code implementations4 May 2016 Michael B. Giles, Mateusz B. Majka, Lukasz Szpruch, Sebastian Vollmer, Konstantinos Zygalakis

We show that this is the first stochastic gradient MCMC method with complexity $\mathcal{O}(\varepsilon^{-2}|\log {\varepsilon}|^{3})$, in contrast to the complexity $\mathcal{O}(\varepsilon^{-3})$ of currently available methods.

Distributed Bayesian Learning with Stochastic Natural-gradient Expectation Propagation and the Posterior Server

no code implementations31 Dec 2015 Leonard Hasenclever, Stefan Webb, Thibaut Lienart, Sebastian Vollmer, Balaji Lakshminarayanan, Charles Blundell, Yee Whye Teh

The posterior server allows scalable and robust Bayesian learning in cases where a data set is stored in a distributed manner across a cluster, with each compute node containing a disjoint subset of data.

Variational Inference

Consistency and fluctuations for stochastic gradient Langevin dynamics

no code implementations1 Sep 2014 Yee Whye Teh, Alexandre Thiéry, Sebastian Vollmer

Applying standard Markov chain Monte Carlo (MCMC) algorithms to large data sets is computationally expensive.

Cannot find the paper you are looking for? You can Submit a new open access paper.