Search Results for author: Aleksandr Beznosikov

Found 25 papers, 4 papers with code

Activations and Gradients Compression for Model-Parallel Training

1 code implementation15 Jan 2024 Mikhail Rudakov, Aleksandr Beznosikov, Yaroslav Kholodov, Alexander Gasnikov

We analyze compression methods such as quantization and TopK compression, and also experiment with error compensation techniques.

Image Classification Language Modelling +1

Ito Diffusion Approximation of Universal Ito Chains for Sampling, Optimization and Boosting

no code implementations9 Oct 2023 Aleksei Ustimenko, Aleksandr Beznosikov

In this work, we consider rather general and broad class of Markov chains, Ito chains, that look like Euler-Maryama discretization of some Stochastic Differential Equation.

Sarah Frank-Wolfe: Methods for Constrained Optimization with Best Rates and Practical Features

no code implementations23 Apr 2023 Aleksandr Beznosikov, David Dobre, Gauthier Gidel

Moreover, our second approach does not require either large batches or full deterministic gradients, which is a typical weakness of many techniques for finite-sum problems.

Similarity, Compression and Local Steps: Three Pillars of Efficient Communications for Distributed Variational Inequalities

no code implementations NeurIPS 2023 Aleksandr Beznosikov, Martin Takáč, Alexander Gasnikov

The methods presented in this paper have the best theoretical guarantees of communication complexity and are significantly ahead of other methods for distributed variational inequalities.

Distributed Computing

SARAH-based Variance-reduced Algorithm for Stochastic Finite-sum Cocoercive Variational Inequalities

no code implementations12 Oct 2022 Aleksandr Beznosikov, Alexander Gasnikov

In this paper we consider the problem of stochastic finite-sum cocoercive variational inequalities.

Smooth Monotone Stochastic Variational Inequalities and Saddle Point Problems: A Survey

no code implementations29 Aug 2022 Aleksandr Beznosikov, Boris Polyak, Eduard Gorbunov, Dmitry Kovalev, Alexander Gasnikov

This paper is a survey of methods for solving smooth (strongly) monotone stochastic variational inequalities.

On Scaled Methods for Saddle Point Problems

no code implementations16 Jun 2022 Aleksandr Beznosikov, Aibek Alanov, Dmitry Kovalev, Martin Takáč, Alexander Gasnikov

Methods with adaptive scaling of different features play a key role in solving saddle point problems, primarily due to Adam's popularity for solving adversarial machine learning problems, including GANS training.

Stochastic Gradient Methods with Preconditioned Updates

no code implementations1 Jun 2022 Abdurakhmon Sadiev, Aleksandr Beznosikov, Abdulla Jasem Almansoori, Dmitry Kamzolov, Rachael Tappenden, Martin Takáč

There are several algorithms for such problems, but existing methods often work poorly when the problem is badly scaled and/or ill-conditioned, and a primary goal of this work is to introduce methods that alleviate this issue.

Optimal Gradient Sliding and its Application to Distributed Optimization Under Similarity

no code implementations30 May 2022 Dmitry Kovalev, Aleksandr Beznosikov, Ekaterina Borodich, Alexander Gasnikov, Gesualdo Scutari

Finally the method is extended to distributed saddle-problems (under function similarity) by means of solving a class of variational inequalities, achieving lower communication and computation complexity bounds.

Distributed Optimization

Stochastic Gradient Descent-Ascent: Unified Theory and New Efficient Methods

1 code implementation15 Feb 2022 Aleksandr Beznosikov, Eduard Gorbunov, Hugo Berard, Nicolas Loizou

Although variants of the new methods are known for solving minimization problems, they were never considered or analyzed for solving min-max problems and VIPs.

Optimal Algorithms for Decentralized Stochastic Variational Inequalities

no code implementations6 Feb 2022 Dmitry Kovalev, Aleksandr Beznosikov, Abdurakhmon Sadiev, Michael Persiianov, Peter Richtárik, Alexander Gasnikov

Our algorithms are the best among the available literature not only in the decentralized stochastic case, but also in the decentralized deterministic and non-distributed stochastic cases.

Distributed Saddle-Point Problems Under Data Similarity

no code implementations NeurIPS 2021 Aleksandr Beznosikov, Gesualdo Scutari, Alexander Rogozin, Alexander Gasnikov

We study solution methods for (strongly-)convex-(strongly)-concave Saddle-Point Problems (SPPs) over networks of two type--master/workers (thus centralized) architectures and mesh (thus decentralized) networks.

Random-reshuffled SARAH does not need a full gradient computations

no code implementations26 Nov 2021 Aleksandr Beznosikov, Martin Takáč

The StochAstic Recursive grAdient algoritHm (SARAH) algorithm is a variance reduced variant of the Stochastic Gradient Descent (SGD) algorithm that needs a gradient of the objective function from time to time.

Distributed Methods with Compressed Communication for Solving Variational Inequalities, with Theoretical Guarantees

no code implementations7 Oct 2021 Aleksandr Beznosikov, Peter Richtárik, Michael Diskin, Max Ryabinin, Alexander Gasnikov

Due to these considerations, it is important to equip existing methods with strategies that would allow to reduce the volume of transmitted information during training while obtaining a model of comparable quality.

Distributed Computing Federated Learning

Distributed Saddle-Point Problems Under Similarity

1 code implementation22 Jul 2021 Aleksandr Beznosikov, Gesualdo Scutari, Alexander Rogozin, Alexander Gasnikov

We study solution methods for (strongly-)convex-(strongly)-concave Saddle-Point Problems (SPPs) over networks of two type - master/workers (thus centralized) architectures and meshed (thus decentralized) networks.

Decentralized Local Stochastic Extra-Gradient for Variational Inequalities

no code implementations15 Jun 2021 Aleksandr Beznosikov, Pavel Dvurechensky, Anastasia Koloskova, Valentin Samokhin, Sebastian U Stich, Alexander Gasnikov

We extend the stochastic extragradient method to this very general setting and theoretically analyze its convergence rate in the strongly-monotone, monotone, and non-monotone (when a Minty solution exists) settings.

Federated Learning

Decentralized Personalized Federated Learning for Min-Max Problems

no code implementations14 Jun 2021 Ekaterina Borodich, Aleksandr Beznosikov, Abdurakhmon Sadiev, Vadim Sushko, Nikolay Savelyev, Martin Takáč, Alexander Gasnikov

Personalized Federated Learning (PFL) has witnessed remarkable advancements, enabling the development of innovative machine learning applications that preserve the privacy of training data.

Distributed Optimization Personalized Federated Learning

Zeroth-Order Algorithms for Smooth Saddle-Point Problems

no code implementations21 Sep 2020 Abdurakhmon Sadiev, Aleksandr Beznosikov, Pavel Dvurechensky, Alexander Gasnikov

In particular, our analysis shows that in the case when the feasible set is a direct product of two simplices, our convergence rate for the stochastic term is only by a $\log n$ factor worse than for the first-order methods.

Gradient-Free Methods for Saddle-Point Problem

no code implementations12 May 2020 Aleksandr Beznosikov, Abdurakhmon Sadiev, Alexander Gasnikov

In the second part of the paper, we analyze the case when such an assumption cannot be made, we propose a general approach on how to modernize the method to solve this problem, and also we apply this approach to particular cases of some classical sets.

On Biased Compression for Distributed Learning

no code implementations27 Feb 2020 Aleksandr Beznosikov, Samuel Horváth, Peter Richtárik, Mher Safaryan

In the last few years, various communication compression techniques have emerged as an indispensable tool helping to alleviate the communication bottleneck in distributed learning.

Cannot find the paper you are looking for? You can Submit a new open access paper.