Search Results for author: Tomohiro Hayase

Found 8 papers, 1 papers with code

MLP-Mixer as a Wide and Sparse MLP

no code implementations2 Jun 2023 Tomohiro Hayase, Ryo Karakida

Multi-layer perceptron (MLP) is a fundamental component of deep learning that has been extensively employed for various problems.

Understanding Gradient Regularization in Deep Learning: Efficient Finite-Difference Computation and Implicit Bias

no code implementations6 Oct 2022 Ryo Karakida, Tomoumi Takase, Tomohiro Hayase, Kazuki Osawa

In this study, we first reveal that a specific finite-difference computation, composed of both gradient ascent and descent steps, reduces the computational cost of GR.

Asymptotic Freeness of Layerwise Jacobians Caused by Invariance of Multilayer Perceptron: The Haar Orthogonal Case

no code implementations24 Mar 2021 Benoit Collins, Tomohiro Hayase

Lastly, we can replace each weight with a Haar orthogonal random matrix independent of the Jacobian of the activation function using this key fact.

Layer-Wise Interpretation of Deep Neural Networks Using Identity Initialization

no code implementations26 Feb 2021 Shohei Kubota, Hideaki Hayashi, Tomohiro Hayase, Seiichi Uchida

The interpretability of neural networks (NNs) is a challenging but essential topic for transparency in the decision-making process using machine learning.

Classification Decision Making +1

Selective Forgetting of Deep Networks at a Finer Level than Samples

no code implementations22 Dec 2020 Tomohiro Hayase, Suguru Yasutomi, Takashi Katoh

Such forgetting is crucial also in a practical sense since the deployed DNNs may be trained on the data with outliers, poisoned by attackers, or with leaked/sensitive information.

Continual Learning

The Spectrum of Fisher Information of Deep Networks Achieving Dynamical Isometry

no code implementations14 Jun 2020 Tomohiro Hayase, Ryo Karakida

We investigate the spectral distribution of the conditional FIM, which is the FIM given a single sample, by focusing on fully-connected networks achieving dynamical isometry.

Almost Sure Asymptotic Freeness of Neural Network Jacobian with Orthogonal Weights

no code implementations11 Aug 2019 Tomohiro Hayase

A well-conditioned Jacobian spectrum has a vital role in preventing exploding or vanishing gradients and speeding up learning of deep neural networks.

Cannot find the paper you are looking for? You can Submit a new open access paper.