Search Results for author: Johannes Schmidt-Hieber

Found 12 papers, 1 papers with code

Convergence guarantees for forward gradient descent in the linear regression model

no code implementations26 Sep 2023 Thijs Bos, Johannes Schmidt-Hieber

Renewed interest in the relationship between artificial and biological neural networks motivates the study of gradient-free methods.

regression

Hebbian learning inspired estimation of the linear regression parameters from queries

no code implementations26 Sep 2023 Johannes Schmidt-Hieber, Wouter M Koolen

In this work, we study a variation of this Hebbian learning rule to recover the regression vector in the linear regression model.

regression

Dropout Regularization Versus $\ell_2$-Penalization in the Linear Model

no code implementations18 Jun 2023 Gabriel Clara, Sophie Langer, Johannes Schmidt-Hieber

We investigate the statistical behavior of gradient descent iterates with dropout in the linear regression model.

L2 Regularization regression

Interpreting learning in biological neural networks as zero-order optimization method

no code implementations27 Jan 2023 Johannes Schmidt-Hieber

In particular, the locality in the updating rule of the connection parameters in biological neural networks (BNNs) makes it biologically implausible that the learning of the brain is based on gradient descent.

On the inability of Gaussian process regression to optimally learn compositional functions

no code implementations16 May 2022 Matteo Giordano, Kolyan Ray, Johannes Schmidt-Hieber

We rigorously prove that deep Gaussian process priors can outperform Gaussian process priors if the target function has a compositional structure.

regression

On generalization bounds for deep networks based on loss surface implicit regularization

1 code implementation12 Jan 2022 Masaaki Imaizumi, Johannes Schmidt-Hieber

We argue that under reasonable assumptions, the local geometry forces SGD to stay close to a low dimensional subspace and that this induces another form of implicit regularization and results in tighter bounds on the generalization error for deep neural networks.

Generalization Bounds Learning Theory

Convergence rates of deep ReLU networks for multiclass classification

no code implementations2 Aug 2021 Thijs Bos, Johannes Schmidt-Hieber

For classification problems, trained deep neural networks return probabilities of class memberships.

Classification

The Kolmogorov-Arnold representation theorem revisited

no code implementations31 Jul 2020 Johannes Schmidt-Hieber

There is a longstanding debate whether the Kolmogorov-Arnold representation theorem can explain the use of more than one hidden layer in neural networks.

On lower bounds for the bias-variance trade-off

no code implementations30 May 2020 Alexis Derumigny, Johannes Schmidt-Hieber

In a second part of the article, the abstract lower bounds are applied to several statistical models including the Gaussian white noise model, a boundary estimation problem, the Gaussian sequence model and the high-dimensional linear regression model.

Deep ReLU network approximation of functions on a manifold

no code implementations2 Aug 2019 Johannes Schmidt-Hieber

Whereas recovery of the manifold from data is a well-studied topic, approximation rates for functions defined on manifolds are less known.

regression

Nonparametric regression using deep neural networks with ReLU activation function

no code implementations22 Aug 2017 Johannes Schmidt-Hieber

It is shown that estimators based on sparsely connected deep neural networks with ReLU activation function and properly chosen network architecture achieve the minimax rates of convergence (up to $\log n$-factors) under a general composition assumption on the regression function.

Additive models regression

Cannot find the paper you are looking for? You can Submit a new open access paper.