no code implementations • ICML 2020 • Rotem Mulayoff, Tomer Michaeli
In this paper, we characterize the wide minima in linear neural networks trained with a quadratic loss.
no code implementations • 21 Feb 2024 • Amitay Bar, Rotem Mulayoff, Tomer Michaeli, Ronen Talmon
Langevin dynamics (LD) is widely used for sampling from distributions and for optimization.
no code implementations • 30 Jun 2023 • Mor Shpigel Nacson, Rotem Mulayoff, Greg Ongie, Tomer Michaeli, Daniel Soudry
Finally, we prove that if a function is sufficiently smooth (in a Sobolev sense) then it can be approximated arbitrarily well using shallow ReLU networks that correspond to stable solutions of gradient descent.
no code implementations • 13 Jun 2023 • Rotem Mulayoff, Tomer Michaeli
Furthermore, we show that SGD's stability threshold is equivalent to that of a mixture process which takes in each iteration a full batch gradient step w. p.
1 code implementation • 20 Mar 2023 • René Haas, Inbar Huberman-Spiegelglas, Rotem Mulayoff, Tomer Michaeli
Recently, a semantic latent space for DDMs, coined `$h$-space', was shown to facilitate semantic image editing in a way reminiscent of GANs.
no code implementations • NeurIPS 2021 • Rotem Mulayoff, Tomer Michaeli, Daniel Soudry
First, we extend the existing knowledge on minima stability to non-differentiable minima, which are common in ReLU nets.
no code implementations • 9 Apr 2020 • Felix Dietrich, Or Yair, Rotem Mulayoff, Ronen Talmon, Ioannis G. Kevrekidis
We show analytically that our method is guaranteed to provide a set of orthogonal functions that are as jointly smooth as possible, ordered by increasing Dirichlet energy from the smoothest to the least smooth.
no code implementations • 11 Feb 2020 • Rotem Mulayoff, Tomer Michaeli
In this paper, we characterize the flat minima in linear neural networks trained with a quadratic loss.
no code implementations • ICML 2018 • Andrey Zhitnikov, Rotem Mulayoff, Tomer Michaeli
In many areas of neuroscience and biological data analysis, it is desired to reveal common patterns among a group of subjects.