Search Results for author: Nidham Gazagnadou

Found 5 papers, 2 papers with code

FedP3: Federated Personalized and Privacy-friendly Network Pruning under Model Heterogeneity

no code implementations15 Apr 2024 Kai Yi, Nidham Gazagnadou, Peter Richtárik, Lingjuan Lyu

The interest in federated learning has surged in recent research due to its unique ability to train a global model using privacy-secured information held locally on each client.

Federated Learning Network Pruning

On the Hardness of Robustness Transfer: A Perspective from Rademacher Complexity over Symmetric Difference Hypothesis Space

no code implementations23 Feb 2023 Yuyang Deng, Nidham Gazagnadou, Junyuan Hong, Mehrdad Mahdavi, Lingjuan Lyu

Recent studies demonstrated that the adversarially robust learning under $\ell_\infty$ attack is harder to generalize to different domains than standard domain adaptation.

Binary Classification Domain Generalization +1

Cutting Some Slack for SGD with Adaptive Polyak Stepsizes

no code implementations24 Feb 2022 Robert M. Gower, Mathieu Blondel, Nidham Gazagnadou, Fabian Pedregosa

We use this insight to develop new variants of the SPS method that are better suited to nonlinear models.

Towards closing the gap between the theory and practice of SVRG

1 code implementation NeurIPS 2019 Othmane Sebbouh, Nidham Gazagnadou, Samy Jelassi, Francis Bach, Robert M. Gower

Among the very first variance reduced stochastic methods for solving the empirical risk minimization problem was the SVRG method (Johnson & Zhang 2013).

Optimal mini-batch and step sizes for SAGA

2 code implementations31 Jan 2019 Nidham Gazagnadou, Robert M. Gower, Joseph Salmon

Using these bounds, and since the SAGA algorithm is part of this JacSketch family, we suggest a new standard practice for setting the step sizes and mini-batch size for SAGA that are competitive with a numerical grid search.

Cannot find the paper you are looking for? You can Submit a new open access paper.