Search Results for author: Giorgio Patrini

Found 12 papers, 3 papers with code

SEALion: a Framework for Neural Network Inference on Encrypted Data

no code implementations29 Apr 2019 Tim van Elsloo, Giorgio Patrini, Hamish Ivey-Law

We present SEALion: an extensible framework for privacy-preserving machine learning with homomorphic encryption.

Privacy Preserving Transfer Learning

Three Tools for Practical Differential Privacy

no code implementations7 Dec 2018 Koen Lennart van der Veen, Ruben Seggers, Peter Bloem, Giorgio Patrini

Differentially private learning on real-world data poses challenges for standard machine learning practice: privacy guarantees are difficult to interpret, hyperparameter tuning on private data reduces the privacy budget, and ad-hoc privacy attacks are often required to test model privacy.

BIG-bench Machine Learning

Sinkhorn AutoEncoders

2 code implementations ICLR 2019 Giorgio Patrini, Rianne van den Berg, Patrick Forré, Marcello Carioni, Samarth Bhargav, Max Welling, Tim Genewein, Frank Nielsen

We show that minimizing the p-Wasserstein distance between the generator and the true data distribution is equivalent to the unconstrained min-min optimization of the p-Wasserstein distance between the encoder aggregated posterior and the prior in latent space, plus a reconstruction error.

Probabilistic Programming

Entity Resolution and Federated Learning get a Federated Resolution

no code implementations11 Mar 2018 Richard Nock, Stephen Hardy, Wilko Henecka, Hamish Ivey-Law, Giorgio Patrini, Guillaume Smith, Brian Thorne

In our experiments, we modify a simple token-based entity resolution algorithm so that it indeed aims at avoiding matching rows belonging to different classes, and perform experiments in the setting where entity resolution relies on noisy data, which is very relevant to real world domains.

Entity Resolution Federated Learning +1

Private federated learning on vertically partitioned data via entity resolution and additively homomorphic encryption

no code implementations29 Nov 2017 Stephen Hardy, Wilko Henecka, Hamish Ivey-Law, Richard Nock, Giorgio Patrini, Guillaume Smith, Brian Thorne

Our results bring a clear and strong support for federated learning: under reasonable assumptions on the number and magnitude of entity resolution's mistakes, it can be extremely beneficial to carry out federated learning in the setting where each peer's data provides a significant uplift to the other.

Entity Resolution Federated Learning +1

Tsallis Regularized Optimal Transport and Ecological Inference

1 code implementation15 Sep 2016 Boris Muzellec, Richard Nock, Giorgio Patrini, Frank Nielsen

We also present the first application of optimal transport to the problem of ecological inference, that is, the reconstruction of joint distributions from their marginals, a problem of large interest in the social sciences.

Making Deep Neural Networks Robust to Label Noise: a Loss Correction Approach

2 code implementations CVPR 2017 Giorgio Patrini, Alessandro Rozza, Aditya Menon, Richard Nock, Lizhen Qu

We present a theoretically grounded approach to train deep neural networks, including recurrent networks, subject to class-dependent label noise.

Ranked #2 on Image Classification on Clothing1M (using clean data) (using extra training data)

Learning with noisy labels Noise Estimation

Fast Learning from Distributed Datasets without Entity Matching

no code implementations13 Mar 2016 Giorgio Patrini, Richard Nock, Stephen Hardy, Tiberio Caetano

Our goal is to learn a classifier in the cross product space of the two domains, in the hard case in which no shared ID is available -- e. g. due to anonymization.

Entity Resolution

Loss factorization, weakly supervised learning and label noise robustness

no code implementations8 Feb 2016 Giorgio Patrini, Frank Nielsen, Richard Nock, Marcello Carioni

We prove that the empirical risk of most well-known loss functions factors into a linear term aggregating all labels with a term that is label free, and can further be expressed by sums of the loss.

Generalization Bounds Weakly-supervised Learning

Rademacher Observations, Private Data, and Boosting

no code implementations9 Feb 2015 Richard Nock, Giorgio Patrini, Arik Friedman

We show that rados comply with various privacy requirements that make them good candidates for machine learning in a privacy framework.

(Almost) No Label No Cry

no code implementations NeurIPS 2014 Giorgio Patrini, Richard Nock, Paul Rivera, Tiberio Caetano

In Learning with Label Proportions (LLP), the objective is to learn a supervised classifier when, instead of labels, only label proportions for bags of observations are known.

Generalization Bounds Privacy Preserving +1

Cannot find the paper you are looking for? You can Submit a new open access paper.