1 code implementation • 12 Feb 2024 • David Selby, Kai Spriestersbach, Yuichiro Iwashita, Dennis Bappert, Archana Warrier, Sumantrak Mukherjee, Muhammad Nabeel Asim, Koichi Kise, Sebastian Vollmer
Large language models (LLMs) have been extensively studied for their abilities to generate convincing natural language sequences, however their utility for quantitative information retrieval is less well understood.
no code implementations • 16 Jan 2024 • Rahul Sharma, Sergey Redyuk, Sumantrak Mukherjee, Andrea Sipka, Sebastian Vollmer, David Selby
Explainable AI (XAI) and interpretable machine learning methods help to build trust in model predictions and derived insights, yet also present a perverse incentive for analysts to manipulate XAI metrics to support pre-specified conclusions.
1 code implementation • 26 May 2022 • Raphael Sonabend, Florian Pfisterer, Alan Mishler, Moritz Schauer, Lukas Burk, Sumantrak Mukherjee, Sebastian Vollmer
In this paper we explore how to utilise existing survival metrics to measure bias with group fairness metrics.
1 code implementation • 4 Feb 2022 • Jen Ning Lim, Sebastian Vollmer, Lorenz Wolf, Andrew Duncan
Their ability to incorporate domain-specific choices and constraints into the structure of the model through composition make EBMs an appealing candidate for applications in physics, biology and computer vision and various other fields.
1 code implementation • 9 Dec 2021 • Raphael Sonabend, Andreas Bender, Sebastian Vollmer
In this paper we consider how to evaluate survival distribution predictions with measures of discrimination.
no code implementations • 24 Aug 2021 • Sahra Ghalebikesabi, Harrison Wilde, Jack Jewson, Arnaud Doucet, Sebastian Vollmer, Chris Holmes
Increasing interest in privacy-preserving machine learning has led to new and evolved approaches for generating private synthetic data from undisclosed real data.
no code implementations • 16 Nov 2020 • Harrison Wilde, Jack Jewson, Sebastian Vollmer, Chris Holmes
There is significant growth and interest in the use of synthetic data as an enabler for machine learning in environments where the release of real data is restricted due to privacy or availability constraints.
no code implementations • 4 Nov 2020 • Ashrya Agrawal, Florian Pfisterer, Bernd Bischl, Francois Buet-Golfouse, Srijan Sood, Jiahao Chen, Sameena Shah, Sebastian Vollmer
We present an empirical study of debiasing methods for classifiers, showing that debiasers often fail in practice to generalize out-of-sample, and can in fact make fairness worse rather than better.
1 code implementation • 30 Jul 2020 • Harrison Wilde, Lucia Lushi Chen, Austin Nguyen, Zoe Kimpel, Joshua Sidgwick, Adolfo De Unanue, Davide Veronese, Bilal Mateen, Rayid Ghani, Sebastian Vollmer
Rough sleeping is a chronic problem faced by some of the most disadvantaged people in modern society.
no code implementations • 21 Dec 2018 • Sebastian Vollmer, Bilal A. Mateen, Gergo Bohner, Franz J. Király, Rayid Ghani, Pall Jonsson, Sarah Cumbers, Adrian Jonas, Katherine S. L. McAllister, Puja Myles, David Granger, Mark Birse, Richard Branson, Karel GM Moons, Gary S Collins, John P. A. Ioannidis, Chris Holmes, Harry Hemingway
Machine learning (ML), artificial intelligence (AI) and other modern statistical methods are providing new opportunities to operationalize previously untapped and rapidly growing sources of data for patient benefit.
no code implementations • 15 Sep 2016 • Mike Giles, Tigran Nagapetyan, Lukasz Szpruch, Sebastian Vollmer, Konstantinos Zygalakis
In contrast to MCMC methods, Stochastic Gradient MCMC (SGMCMC) algorithms such as the Stochastic Gradient Langevin Dynamics (SGLD) only require access to a batch of the data set at every step.
no code implementations • 4 May 2016 • Michael B. Giles, Mateusz B. Majka, Lukasz Szpruch, Sebastian Vollmer, Konstantinos Zygalakis
We show that this is the first stochastic gradient MCMC method with complexity $\mathcal{O}(\varepsilon^{-2}|\log {\varepsilon}|^{3})$, in contrast to the complexity $\mathcal{O}(\varepsilon^{-3})$ of currently available methods.
no code implementations • 31 Dec 2015 • Leonard Hasenclever, Stefan Webb, Thibaut Lienart, Sebastian Vollmer, Balaji Lakshminarayanan, Charles Blundell, Yee Whye Teh
The posterior server allows scalable and robust Bayesian learning in cases where a data set is stored in a distributed manner across a cluster, with each compute node containing a disjoint subset of data.
no code implementations • 1 Sep 2014 • Yee Whye Teh, Alexandre Thiéry, Sebastian Vollmer
Applying standard Markov chain Monte Carlo (MCMC) algorithms to large data sets is computationally expensive.