Search Results for author: Vitaly Feldman

Found 56 papers, 4 papers with code

Private Vector Mean Estimation in the Shuffle Model: Optimal Rates Require Many Messages

no code implementations16 Apr 2024 Hilal Asi, Vitaly Feldman, Jelani Nelson, Huy L. Nguyen, Kunal Talwar, Samson Zhou

We study the problem of private vector mean estimation in the shuffle model of privacy where $n$ users each have a unit vector $v^{(i)} \in\mathbb{R}^d$.

Faster Convergence with Multiway Preferences

no code implementations19 Dec 2023 Aadirupa Saha, Vitaly Feldman, Tomer Koren, Yishay Mansour

We next study a $m$-multiway comparison (`battling') feedback, where the learner can get to see the argmin feedback of $m$-subset of queried points and show a convergence rate of $\smash{\widetilde O}(\frac{d}{ \min\{\log m, d\}\epsilon })$.

Mean Estimation with User-level Privacy under Data Heterogeneity

no code implementations28 Jul 2023 Rachel Cummings, Vitaly Feldman, Audra McMillan, Kunal Talwar

In this work we propose a simple model of heterogeneous user data that allows user data to differ in both distribution and quantity of data, and provide a method for estimating the population-level mean while preserving user-level differential privacy.

Differentially Private Heavy Hitter Detection using Federated Analytics

no code implementations21 Jul 2023 Karan Chadha, Junye Chen, John Duchi, Vitaly Feldman, Hanieh Hashemi, Omid Javidbakht, Audra McMillan, Kunal Talwar

In this work, we study practical heuristics to improve the performance of prefix-tree based algorithms for differentially private heavy hitter detection.

Near-Optimal Algorithms for Private Online Optimization in the Realizable Regime

no code implementations27 Feb 2023 Hilal Asi, Vitaly Feldman, Tomer Koren, Kunal Talwar

We also develop an adaptive algorithm for the small-loss setting with regret $O(L^\star\log d + \varepsilon^{-1} \log^{1. 5}{d})$ where $L^\star$ is the total loss of the best expert.

Subspace Recovery from Heterogeneous Data with Non-isotropic Noise

no code implementations24 Oct 2022 John Duchi, Vitaly Feldman, Lunjia Hu, Kunal Talwar

Our goal is to recover the linear subspace shared by $\mu_1,\ldots,\mu_n$ using the data points from all users, where every data point from user $i$ is formed by adding an independent mean-zero noise vector to $\mu_i$.

Federated Learning

Private Online Prediction from Experts: Separations and Faster Rates

no code implementations24 Oct 2022 Hilal Asi, Vitaly Feldman, Tomer Koren, Kunal Talwar

Our lower bounds also show a separation between pure and approximate differential privacy for adaptive adversaries where the latter is necessary to achieve the non-private $O(\sqrt{T})$ regret.

Optimal Algorithms for Mean Estimation under Local Differential Privacy

no code implementations5 May 2022 Hilal Asi, Vitaly Feldman, Kunal Talwar

We show that PrivUnit (Bhowmick et al. 2018) with optimized parameters achieves the optimal variance among a large family of locally private randomizers.

Private Frequency Estimation via Projective Geometry

1 code implementation1 Mar 2022 Vitaly Feldman, Jelani Nelson, Huy Lê Nguyen, Kunal Talwar

In many parameter settings used in practice this is a significant improvement over the $ O(n+k^2)$ computation cost that is achieved by the recent PI-RAPPOR algorithm (Feldman and Talwar; 2021).

Individual Privacy Accounting via a Rényi Filter

no code implementations NeurIPS 2021 Vitaly Feldman, Tijana Zrnic

In this work, we give a method for tighter privacy loss accounting based on the value of a personalized privacy loss estimate for each individual in each analysis.

Private Stochastic Convex Optimization: Optimal Rates in $\ell_1$ Geometry

no code implementations2 Mar 2021 Hilal Asi, Vitaly Feldman, Tomer Koren, Kunal Talwar

Stochastic convex optimization over an $\ell_1$-bounded domain is ubiquitous in machine learning applications such as LASSO but remains poorly understood when learning with differential privacy.

Lossless Compression of Efficient Private Local Randomizers

no code implementations24 Feb 2021 Vitaly Feldman, Kunal Talwar

Here we demonstrate a general approach that, under standard cryptographic assumptions, compresses every efficient LDP algorithm with negligible loss in privacy and utility guarantees.

Hiding Among the Clones: A Simple and Nearly Optimal Analysis of Privacy Amplification by Shuffling

1 code implementation23 Dec 2020 Vitaly Feldman, Audra McMillan, Kunal Talwar

As a direct corollary of our analysis we derive a simple and nearly optimal algorithm for frequency estimation in the shuffle model of privacy.

When is Memorization of Irrelevant Training Data Necessary for High-Accuracy Learning?

1 code implementation11 Dec 2020 Gavin Brown, Mark Bun, Vitaly Feldman, Adam Smith, Kunal Talwar

Our problems are simple and fairly natural variants of the next-symbol prediction and the cluster labeling tasks.

Memorization

Individual Privacy Accounting via a Renyi Filter

no code implementations NeurIPS 2021 Vitaly Feldman, Tijana Zrnic

We consider a sequential setting in which a single dataset of individuals is used to perform adaptively-chosen analyses, while ensuring that the differential privacy loss of each participant does not exceed a pre-specified privacy budget.

What Neural Networks Memorize and Why: Discovering the Long Tail via Influence Estimation

no code implementations NeurIPS 2020 Vitaly Feldman, Chiyuan Zhang

First, natural image and data distributions are (informally) known to be long-tailed, that is have a significant fraction of rare and atypical examples.

Memorization

Private Stochastic Convex Optimization: Optimal Rates in Linear Time

no code implementations10 May 2020 Vitaly Feldman, Tomer Koren, Kunal Talwar

We also give a linear-time algorithm achieving the optimal bound on the excess loss for the strongly convex case, as well as a faster algorithm for the non-smooth case.

PAC learning with stable and private predictions

no code implementations24 Nov 2019 Yuval Dagan, Vitaly Feldman

For $\epsilon$-differentially private prediction we give two new algorithms: one using $\tilde O(d/(\alpha^2\epsilon))$ samples and another one using $\tilde O(d^2/(\alpha\epsilon) + d/\alpha^2)$ samples.

Binary Classification PAC learning

Interaction is necessary for distributed learning with privacy or communication constraints

no code implementations11 Nov 2019 Yuval Dagan, Vitaly Feldman

Our main result is an exponential lower bound on the number of samples necessary to solve the standard task of learning a large-margin linear separator in the non-interactive LDP model.

Private Stochastic Convex Optimization with Optimal Rates

no code implementations NeurIPS 2019 Raef Bassily, Vitaly Feldman, Kunal Talwar, Abhradeep Thakurta

A long line of existing work on private convex optimization focuses on the empirical loss and derives asymptotically tight bounds on the excess empirical loss.

Does Learning Require Memorization? A Short Tale about a Long Tail

no code implementations12 Jun 2019 Vitaly Feldman

In our model, data is sampled from a mixture of subpopulations and our results show that memorization is necessary whenever the distribution of subpopulation frequencies is long-tailed.

Memorization Model Compression

The advantages of multiple classes for reducing overfitting from test set reuse

no code implementations24 May 2019 Vitaly Feldman, Roy Frostig, Moritz Hardt

We show a new upper bound of $\tilde O(\max\{\sqrt{k\log(n)/(mn)}, k/n\})$ on the worst-case bias that any attack can achieve in a prediction problem with $m$ classes.

Binary Classification

High probability generalization bounds for uniformly stable algorithms with nearly optimal rate

no code implementations27 Feb 2019 Vitaly Feldman, Jan Vondrak

Specifically, their bound on the estimation error of any $\gamma$-uniformly stable learning algorithm on $n$ samples and range in $[0, 1]$ is $O(\gamma \sqrt{n \log(1/\delta)} + \sqrt{\log(1/\delta)/n})$ with probability $\geq 1-\delta$.

Generalization Bounds Vocal Bursts Intensity Prediction

Generalization Bounds for Uniformly Stable Algorithms

no code implementations NeurIPS 2018 Vitaly Feldman, Jan Vondrak

Specifically, for a loss function with range bounded in $[0, 1]$, the generalization error of a $\gamma$-uniformly stable learning algorithm on $n$ samples is known to be within $O((\gamma +1/n) \sqrt{n \log(1/\delta)})$ of the empirical error with probability at least $1-\delta$.

Generalization Bounds

Amplification by Shuffling: From Local to Central Differential Privacy via Anonymity

no code implementations29 Nov 2018 Úlfar Erlingsson, Vitaly Feldman, Ilya Mironov, Ananth Raghunathan, Kunal Talwar, Abhradeep Thakurta

We study the collection of such statistics in the local differential privacy (LDP) model, and describe an algorithm whose privacy cost is polylogarithmic in the number of changes to a user's value.

Locally Private Learning without Interaction Requires Separation

no code implementations NeurIPS 2019 Amit Daniely, Vitaly Feldman

The only lower bound we are aware of is for PAC learning an artificial class of functions with respect to a uniform distribution (Kasiviswanathan et al. 2011).

PAC learning

Privacy Amplification by Iteration

no code implementations20 Aug 2018 Vitaly Feldman, Ilya Mironov, Kunal Talwar, Abhradeep Thakurta

In addition, we demonstrate that we can achieve guarantees similar to those obtainable using the privacy-amplification-by-sampling technique in several natural settings where that technique cannot be applied.

Privacy-preserving Prediction

no code implementations27 Mar 2018 Cynthia Dwork, Vitaly Feldman

We demonstrate that this overhead can be avoided for the well-studied class of thresholds on a line and for a number of standard settings of convex regression.

PAC learning Privacy Preserving +1

The Everlasting Database: Statistical Validity at a Fair Price

no code implementations NeurIPS 2018 Blake Woodworth, Vitaly Feldman, Saharon Rosset, Nathan Srebro

The problem of handling adaptivity in data analysis, intentional or not, permeates a variety of fields, including test-set overfitting in ML challenges and the accumulation of invalid scientific discoveries.

Calibrating Noise to Variance in Adaptive Data Analysis

no code implementations19 Dec 2017 Vitaly Feldman, Thomas Steinke

We demonstrate that a simple and natural algorithm based on adding noise scaled to the standard deviation of the query provides our notion of stability.

Generalization for Adaptively-chosen Estimators via Stable Median

no code implementations15 Jun 2017 Vitaly Feldman, Thomas Steinke

We present an algorithm that estimates the expectations of $k$ arbitrary adaptively-chosen real-valued estimators using a number of samples that scales as $\sqrt{k}$.

On the Power of Learning from $k$-Wise Queries

no code implementations28 Feb 2017 Vitaly Feldman, Badih Ghazi

Hence it is natural to ask whether algorithms using $k$-wise queries can solve learning problems more efficiently and by how much.

PAC learning

Dealing with Range Anxiety in Mean Estimation via Statistical Queries

no code implementations20 Nov 2016 Vitaly Feldman

We give algorithms for estimating the expectation of a given real-valued function $\phi:X\to {\bf R}$ on a sample drawn randomly from some unknown distribution $D$ over domain $X$, namely ${\bf E}_{{\bf x}\sim D}[\phi({\bf x})]$.

Generalization of ERM in Stochastic Convex Optimization: The Dimension Strikes Back

no code implementations NeurIPS 2016 Vitaly Feldman

In stochastic convex optimization the goal is to minimize a convex function $F(x) \doteq {\mathbf E}_{{\mathbf f}\sim D}[{\mathbf f}(x)]$ over a convex set $\cal K \subset {\mathbb R}^d$ where $D$ is some unknown distribution and each $f(\cdot)$ in the support of $D$ is convex over $\cal K$.

A General Characterization of the Statistical Query Complexity

no code implementations7 Aug 2016 Vitaly Feldman

We give applications of our techniques to two open problems in learning theory and to algorithms that are subject to memory and communication constraints.

Learning Theory

Statistical Query Algorithms for Mean Vector Estimation and Stochastic Convex Optimization

no code implementations30 Dec 2015 Vitaly Feldman, Cristobal Guzman, Santosh Vempala

Stochastic convex optimization, where the objective is the expectation of a random convex function, is an important and widely used method with numerous applications in machine learning, statistics, operations research and other areas.

BIG-bench Machine Learning

Tight Bounds on Low-degree Spectral Concentration of Submodular and XOS functions

no code implementations13 Apr 2015 Vitaly Feldman, Jan Vondrak

This improves on previous approaches that all showed an upper bound of $O(1/\epsilon^2)$ for submodular and XOS functions.

Combinatorial Optimization PAC learning

Preserving Statistical Validity in Adaptive Data Analysis

no code implementations10 Nov 2014 Cynthia Dwork, Vitaly Feldman, Moritz Hardt, Toniann Pitassi, Omer Reingold, Aaron Roth

We show that, surprisingly, there is a way to estimate an exponential in $n$ number of expectations accurately even if the functions are chosen adaptively.

Two-sample testing

Agnostic Learning of Disjunctions on Symmetric Distributions

no code implementations27 May 2014 Vitaly Feldman, Pravesh Kothari

This directly gives an agnostic learning algorithm for disjunctions on symmetric distributions that runs in time $n^{O( \log{(1/\epsilon)})}$.

Approximate resilience, monotonicity, and the complexity of agnostic learning

no code implementations21 May 2014 Dana Dachman-Soled, Vitaly Feldman, Li-Yang Tan, Andrew Wan, Karl Wimmer

We study the notion of $\mathit{approximate}$ $\mathit{resilience}$ of Boolean functions, where we say that $f$ is $\alpha$-approximately $d$-resilient if $f$ is $\alpha$-close to a $[-1, 1]$-valued $d$-resilient function in $\ell_1$ distance.

Tight Bounds on $\ell_1$ Approximation and Learning of Self-Bounding Functions

no code implementations18 Apr 2014 Vitaly Feldman, Pravesh Kothari, Jan Vondrák

Previous techniques considered stronger $\ell_2$ approximation and proved nearly tight bounds of $\Theta(1/\epsilon^{2})$ on the degree and $2^{\Theta(1/\epsilon^2)}$ on the number of variables.

Sample Complexity Bounds on Differentially Private Learning via Communication Complexity

no code implementations25 Feb 2014 Vitaly Feldman, David Xiao

Our second contribution and the main tool is an equivalence between the sample complexity of (pure) differentially private learning of a concept class $C$ (or $SCDP(C)$) and the randomized one-way communication complexity of the evaluation problem for concepts from $C$.

PAC learning

Statistical Active Learning Algorithms

no code implementations NeurIPS 2013 Maria-Florina F. Balcan, Vitaly Feldman

We describe a framework for designing efficient active learning algorithms that are tolerant to random classification noise.

Active Learning General Classification

Optimal Bounds on Approximation of Submodular and XOS Functions by Juntas

no code implementations12 Jul 2013 Vitaly Feldman, Jan Vondrak

This is the first algorithm in the PMAC model that over the uniform distribution can achieve a constant approximation factor arbitrarily close to 1 for all submodular functions.

PAC learning

Statistical Active Learning Algorithms for Noise Tolerance and Differential Privacy

no code implementations11 Jul 2013 Maria Florina Balcan, Vitaly Feldman

These results combined with our generic conversion lead to the first computationally-efficient algorithms for actively learning some of these concept classes in the presence of random classification noise that provide exponential improvement in the dependence on the error $\epsilon$ over their passive counterparts.

Active Learning General Classification

Learning Coverage Functions and Private Release of Marginals

no code implementations8 Apr 2013 Vitaly Feldman, Pravesh Kothari

As an application of our learning results, we give simple differentially-private algorithms for releasing monotone conjunction counting queries with low average error.

Representation, Approximation and Learning of Submodular Functions Using Low-rank Decision Trees

no code implementations2 Apr 2013 Vitaly Feldman, Pravesh Kothari, Jan Vondrak

We show that these structural results can be exploited to give an attribute-efficient PAC learning algorithm for submodular functions running in time $\tilde{O}(n^2) \cdot 2^{O(1/\epsilon^{4})}$.

Attribute PAC learning

Learning using Local Membership Queries

no code implementations5 Nov 2012 Pranjal Awasthi, Vitaly Feldman, Varun Kanade

We introduce a new model of membership query (MQ) learning, where the learning algorithm is restricted to query points that are \emph{close} to random examples drawn from the underlying distribution.

Learning DNF Expressions from Fourier Spectrum

no code implementations3 Mar 2012 Vitaly Feldman

This property is crucial for learning of DNF expressions over smoothed product distributions, a learning model introduced by Kalai et al. (2009) and inspired by the seminal smoothed analysis model of Spielman and Teng (2001).

Learning Theory PAC learning

A Complete Characterization of Statistical Query Learning with Applications to Evolvability

no code implementations16 Feb 2010 Vitaly Feldman

The preservation of efficiency is achieved using a new boosting technique and allows us to derive a new approach to the design of evolutionary algorithms in Valiant's (2006) model of evolvability.

Evolutionary Algorithms PAC learning

Cannot find the paper you are looking for? You can Submit a new open access paper.