Search Results for author: Daniil Merkulov

Found 8 papers, 4 papers with code

NAG-GS: Semi-Implicit, Accelerated and Robust Stochastic Optimizer

2 code implementations29 Sep 2022 Valentin Leplat, Daniil Merkulov, Aleksandr Katrutsa, Daniel Bershatsky, Olga Tsymboi, Ivan Oseledets

Classical machine learning models such as deep neural networks are usually trained by using Stochastic Gradient Descent-based (SGD) algorithms.

Memory-Efficient Backpropagation through Large Linear Layers

2 code implementations31 Jan 2022 Daniel Bershatsky, Aleksandr Mikhalev, Alexandr Katrutsa, Julia Gusak, Daniil Merkulov, Ivan Oseledets

Also, we investigate the variance of the gradient estimate induced by the randomized matrix multiplication.

Model Compression

Fast Line Search for Multi-Task Learning

no code implementations2 Oct 2021 Andrey Filatov, Daniil Merkulov

But, usually, line search for the step size is not the best choice due to the large computational time overhead.

Multi-Task Learning

Stochastic gradient algorithms from ODE splitting perspective

no code implementations ICLR Workshop DeepDiffEq 2019 Daniil Merkulov, Ivan Oseledets

We present a different view on stochastic optimization, which goes back to the splitting schemes for approximate solutions of ODE.

regression Stochastic Optimization

Empirical study of extreme overfitting points of neural networks

no code implementations14 Jun 2019 Daniil Merkulov, Ivan Oseledets

In this paper we propose a method of obtaining points of extreme overfitting - parameters of modern neural networks, at which they demonstrate close to 100 % training accuracy, simultaneously with almost zero accuracy on the test sample.

Cannot find the paper you are looking for? You can Submit a new open access paper.