Search Results for author: Rotem Mulayoff

Found 9 papers, 1 papers with code

Unique Properties of Wide Minima in Deep Networks

no code implementations ICML 2020 Rotem Mulayoff, Tomer Michaeli

In this paper, we characterize the wide minima in linear neural networks trained with a quadratic loss.

The Expected Loss of Preconditioned Langevin Dynamics Reveals the Hessian Rank

no code implementations21 Feb 2024 Amitay Bar, Rotem Mulayoff, Tomer Michaeli, Ronen Talmon

Langevin dynamics (LD) is widely used for sampling from distributions and for optimization.

The Implicit Bias of Minima Stability in Multivariate Shallow ReLU Networks

no code implementations30 Jun 2023 Mor Shpigel Nacson, Rotem Mulayoff, Greg Ongie, Tomer Michaeli, Daniel Soudry

Finally, we prove that if a function is sufficiently smooth (in a Sobolev sense) then it can be approximated arbitrarily well using shallow ReLU networks that correspond to stable solutions of gradient descent.

Exact Mean Square Linear Stability Analysis for SGD

no code implementations13 Jun 2023 Rotem Mulayoff, Tomer Michaeli

Furthermore, we show that SGD's stability threshold is equivalent to that of a mixture process which takes in each iteration a full batch gradient step w. p.

Discovering Interpretable Directions in the Semantic Latent Space of Diffusion Models

1 code implementation20 Mar 2023 René Haas, Inbar Huberman-Spiegelglas, Rotem Mulayoff, Tomer Michaeli

Recently, a semantic latent space for DDMs, coined `$h$-space', was shown to facilitate semantic image editing in a way reminiscent of GANs.

Attribute Denoising +1

The Implicit Bias of Minima Stability: A View from Function Space

no code implementations NeurIPS 2021 Rotem Mulayoff, Tomer Michaeli, Daniel Soudry

First, we extend the existing knowledge on minima stability to non-differentiable minima, which are common in ReLU nets.

Spectral Discovery of Jointly Smooth Features for Multimodal Data

no code implementations9 Apr 2020 Felix Dietrich, Or Yair, Rotem Mulayoff, Ronen Talmon, Ioannis G. Kevrekidis

We show analytically that our method is guaranteed to provide a set of orthogonal functions that are as jointly smooth as possible, ordered by increasing Dirichlet energy from the smoothest to the least smooth.

Unique Properties of Flat Minima in Deep Networks

no code implementations11 Feb 2020 Rotem Mulayoff, Tomer Michaeli

In this paper, we characterize the flat minima in linear neural networks trained with a quadratic loss.

Revealing Common Statistical Behaviors in Heterogeneous Populations

no code implementations ICML 2018 Andrey Zhitnikov, Rotem Mulayoff, Tomer Michaeli

In many areas of neuroscience and biological data analysis, it is desired to reveal common patterns among a group of subjects.

Cannot find the paper you are looking for? You can Submit a new open access paper.