Search Results for author: Zohar Ringel

Found 13 papers, 0 papers with code

Towards Understanding Inductive Bias in Transformers: A View From Infinity

no code implementations7 Feb 2024 Itay Lavie, Guy Gur-Ari, Zohar Ringel

We study inductive bias in Transformers in the infinitely over-parameterized Gaussian process limit and argue transformers tend to be biased towards more permutation symmetric functions in sequence space.

Inductive Bias

Droplets of Good Representations: Grokking as a First Order Phase Transition in Two Layer Networks

no code implementations5 Oct 2023 Noa Rubin, Inbar Seroussi, Zohar Ringel

A key property of deep neural networks (DNNs) is their ability to learn new features during training.

Speed Limits for Deep Learning

no code implementations27 Jul 2023 Inbar Seroussi, Alexander A. Alemi, Moritz Helias, Zohar Ringel

State-of-the-art neural networks require extreme computational power to train.

Spectral-Bias and Kernel-Task Alignment in Physically Informed Neural Networks

no code implementations12 Jul 2023 Inbar Seroussi, Asaf Miron, Zohar Ringel

Physically informed neural networks (PINNs) are a promising emerging method for solving differential equations.

GPR

Separation of Scales and a Thermodynamic Description of Feature Learning in Some CNNs

no code implementations31 Dec 2021 Inbar Seroussi, Gadi Naveh, Zohar Ringel

Deep neural networks (DNNs) are powerful tools for compressing and distilling information.

A self consistent theory of Gaussian Processes captures feature learning effects in finite CNNs

no code implementations NeurIPS 2021 Gadi Naveh, Zohar Ringel

Deep neural networks (DNNs) in the infinite width/channel limit have received much attention recently, as they provide a clear analytical window to deep learning via mappings to Gaussian Processes (GPs).

Gaussian Processes

Predicting the Outputs of Finite Networks Trained with Noisy Gradients

no code implementations28 Sep 2020 Gadi Naveh, Oded Ben-David, Haim Sompolinsky, Zohar Ringel

A recent line of works studied wide deep neural networks (DNNs) by approximating them as Gaussian Processes (GPs).

Gaussian Processes Image Classification

Learning Curves for Deep Neural Networks: A field theory perspective

no code implementations25 Sep 2019 Omry Cohen, Or Malka, Zohar Ringel

A series of recent works established a rigorous correspondence between very wide deep neural networks (DNNs), trained in a particular manner, and noiseless Bayesian Inference with a certain Gaussian Process (GP) known as the Neural Tangent Kernel (NTK).

Bayesian Inference

Learning Curves for Deep Neural Networks: A Gaussian Field Theory Perspective

no code implementations12 Jun 2019 Omry Cohen, Or Malka, Zohar Ringel

In the past decade, deep neural networks (DNNs) came to the fore as the leading machine learning algorithms for a variety of tasks.

Bayesian Inference Gaussian Processes

The role of a layer in deep neural networks: a Gaussian Process perspective

no code implementations6 Feb 2019 Oded Ben-David, Zohar Ringel

Leveraging this correspondence, we derive the Deep Gaussian Layer-wise loss functions (DGLs) which, we believe, are the first supervised layer-wise loss functions which are both explicit and competitive in terms of accuracy.

Gaussian Processes

Critical Percolation as a Framework to Analyze the Training of Deep Networks

no code implementations ICLR 2018 Zohar Ringel, Rodrigo de Bem

In this paper we approach two relevant deep learning topics: i) tackling of graph structured input data and ii) a better understanding and analysis of deep networks and related learning algorithms.

General Classification

Mutual Information, Neural Networks and the Renormalization Group

no code implementations20 Apr 2017 Maciej Koch-Janusz, Zohar Ringel

Physical systems differring in their microscopic details often display strikingly similar behaviour when probed at macroscopic scales.

Disordered Systems and Neural Networks Statistical Mechanics Information Theory Machine Learning Information Theory Machine Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.