Search Results for author: Rajiv Khanna

Found 28 papers, 4 papers with code

A Precise Characterization of SGD Stability Using Loss Surface Geometry

no code implementations22 Jan 2024 Gregory Dexter, Borja Ocejo, Sathiya Keerthi, Aman Gupta, Ayan Acharya, Rajiv Khanna

In this paper, we delve deeper into the relationship between linear stability and sharpness.

On Memorization and Privacy Risks of Sharpness Aware Minimization

no code implementations30 Sep 2023 Young In Kim, Pratiksha Agrawal, Johannes O. Royset, Rajiv Khanna

In this work, we dissect these performance gains through the lens of data memorization in overparameterized models.

Memorization

Generalization Guarantees via Algorithm-dependent Rademacher Complexity

no code implementations4 Jul 2023 Sarah Sachs, Tim van Erven, Liam Hodgkinson, Rajiv Khanna, Umut Simsekli

Algorithm- and data-dependent generalization bounds are required to explain the generalization behavior of modern machine learning algorithms.

Generalization Bounds

Feature Space Sketching for Logistic Regression

no code implementations24 Mar 2023 Gregory Dexter, Rajiv Khanna, Jawad Raheel, Petros Drineas

We present novel bounds for coreset construction, feature selection, and dimensionality reduction for logistic regression.

Dimensionality Reduction feature selection +1

mSAM: Micro-Batch-Averaged Sharpness-Aware Minimization

no code implementations19 Feb 2023 Kayhan Behdin, Qingquan Song, Aman Gupta, Sathiya Keerthi, Ayan Acharya, Borja Ocejo, Gregory Dexter, Rajiv Khanna, David Durfee, Rahul Mazumder

Modern deep learning models are over-parameterized, where different optima can result in widely varying generalization performance.

Image Classification

Generalization Bounds using Lower Tail Exponents in Stochastic Optimizers

no code implementations2 Aug 2021 Liam Hodgkinson, Umut Şimşekli, Rajiv Khanna, Michael W. Mahoney

Despite the ubiquitous use of stochastic optimization algorithms in machine learning, the precise impact of these algorithms and their dynamics on generalization performance in realistic non-convex settings is still poorly understood.

Generalization Bounds Stochastic Optimization

LocalNewton: Reducing Communication Bottleneck for Distributed Learning

no code implementations16 May 2021 Vipul Gupta, Avishek Ghosh, Michal Derezinski, Rajiv Khanna, Kannan Ramchandran, Michael Mahoney

To enhance practicability, we devise an adaptive scheme to choose L, and we show that this reduces the number of local iterations in worker machines between two model synchronizations as the training proceeds, successively refining the model quality at the master.

Distributed Optimization

Improved guarantees and a multiple-descent curve for Column Subset Selection and the Nystrom method

no code implementations NeurIPS 2020 Michal Derezinski, Rajiv Khanna, Michael W. Mahoney

The Column Subset Selection Problem (CSSP) and the Nystrom method are among the leading tools for constructing small low-rank approximations of large datasets in machine learning and scientific computing.

Boundary thickness and robustness in learning models

1 code implementation NeurIPS 2020 Yaoqing Yang, Rajiv Khanna, Yaodong Yu, Amir Gholami, Kurt Keutzer, Joseph E. Gonzalez, Kannan Ramchandran, Michael W. Mahoney

Using these observations, we show that noise-augmentation on mixup training further increases boundary thickness, thereby combating vulnerability to various forms of adversarial attacks and OOD transforms.

Adversarial Defense Data Augmentation

Bayesian Coresets: Revisiting the Nonconvex Optimization Perspective

1 code implementation1 Jul 2020 Jacky Y. Zhang, Rajiv Khanna, Anastasios Kyrillidis, Oluwasanmi Koyejo

Bayesian coresets have emerged as a promising approach for implementing scalable Bayesian inference.

Bayesian Inference

Improved guarantees and a multiple-descent curve for Column Subset Selection and the Nyström method

no code implementations21 Feb 2020 Michał Dereziński, Rajiv Khanna, Michael W. Mahoney

The Column Subset Selection Problem (CSSP) and the Nystr\"om method are among the leading tools for constructing small low-rank approximations of large datasets in machine learning and scientific computing.

Learning Sparse Distributions using Iterative Hard Thresholding

no code implementations NeurIPS 2019 Jacky Y. Zhang, Rajiv Khanna, Anastasios Kyrillidis, Oluwasanmi Koyejo

Iterative hard thresholding (IHT) is a projected gradient descent algorithm, known to achieve state of the art performance for a wide range of structured estimation problems, such as sparse inference.

Geometric Rates of Convergence for Kernel-based Sampling Algorithms

no code implementations19 Jul 2019 Rajiv Khanna, Liam Hodgkinson, Michael W. Mahoney

The rate of convergence of weighted kernel herding (WKH) and sequential Bayesian quadrature (SBQ), two kernel-based sampling algorithms for estimating integrals with respect to some target probability measure, is investigated.

Interpreting Black Box Predictions using Fisher Kernels

no code implementations23 Oct 2018 Rajiv Khanna, Been Kim, Joydeep Ghosh, Oluwasanmi Koyejo

Research in both machine learning and psychology suggests that salient examples can help humans to interpret learning models.

Data Summarization

Boosting Black Box Variational Inference

1 code implementation NeurIPS 2018 Francesco Locatello, Gideon Dresdner, Rajiv Khanna, Isabel Valera, Gunnar Rätsch

Finally, we present a stopping criterion drawn from the duality gap in the classic FW analyses and exhaustive experiments to illustrate the usefulness of our theoretical and algorithmic contributions.

Variational Inference

IHT dies hard: Provable accelerated Iterative Hard Thresholding

no code implementations26 Dec 2017 Rajiv Khanna, Anastasios Kyrillidis

We study --both in theory and practice-- the use of momentum motions in classic iterative hard thresholding (IHT) methods.

Boosting Variational Inference: an Optimization Perspective

no code implementations5 Aug 2017 Francesco Locatello, Rajiv Khanna, Joydeep Ghosh, Gunnar Rätsch

Variational inference is a popular technique to approximate a possibly intractable Bayesian posterior with a more tractable one.

Variational Inference

Scalable Greedy Feature Selection via Weak Submodularity

no code implementations8 Mar 2017 Rajiv Khanna, Ethan Elenberg, Alexandros G. Dimakis, Sahand Negahban, Joydeep Ghosh

Furthermore, we show that a bounded submodularity ratio can be used to provide data dependent bounds that can sometimes be tighter also for submodular functions.

feature selection

On Approximation Guarantees for Greedy Low Rank Optimization

no code implementations ICML 2017 Rajiv Khanna, Ethan Elenberg, Alexandros G. Dimakis, Sahand Negahban

We provide new approximation guarantees for greedy low rank matrix estimation under standard assumptions of restricted strong convexity and smoothness.

Combinatorial Optimization

A Unified Optimization View on Generalized Matching Pursuit and Frank-Wolfe

no code implementations21 Feb 2017 Francesco Locatello, Rajiv Khanna, Michael Tschannen, Martin Jaggi

Two of the most fundamental prototypes of greedy optimization are the matching pursuit and Frank-Wolfe algorithms.

Restricted Strong Convexity Implies Weak Submodularity

no code implementations2 Dec 2016 Ethan R. Elenberg, Rajiv Khanna, Alexandros G. Dimakis, Sahand Negahban

Our results extend the work of Das and Kempe (2011) from the setting of linear regression to arbitrary objective functions.

feature selection

Examples are not enough, learn to criticize! Criticism for Interpretability

no code implementations NeurIPS 2016 Been Kim, Rajiv Khanna, Oluwasanmi O. Koyejo

Example-based explanations are widely used in the effort to improve the interpretability of highly complex distributions.

Information Projection and Approximate Inference for Structured Sparse Variables

no code implementations12 Jul 2016 Rajiv Khanna, Joydeep Ghosh, Russell Poldrack, Oluwasanmi Koyejo

Approximate inference via information projection has been recently introduced as a general-purpose approach for efficient probabilistic inference given sparse variables.

Pursuits in Structured Non-Convex Matrix Factorizations

no code implementations12 Feb 2016 Rajiv Khanna, Michael Tschannen, Martin Jaggi

Efficiently representing real world data in a succinct and parsimonious manner is of central importance in many fields.

Towards a Better Understanding of Predict and Count Models

no code implementations6 Nov 2015 S. Sathiya Keerthi, Tobias Schnabel, Rajiv Khanna

In a recent paper, Levy and Goldberg pointed out an interesting connection between prediction-based word embedding models and count models based on pointwise mutual information.

L2 Regularization

On Prior Distributions and Approximate Inference for Structured Variables

no code implementations NeurIPS 2014 Oluwasanmi O. Koyejo, Rajiv Khanna, Joydeep Ghosh, Russell Poldrack

In cases where this projection is intractable, we propose a family of parameterized approximations indexed by subsets of the domain.

Cannot find the paper you are looking for? You can Submit a new open access paper.