Search Results for author: Dongsung Huh

Found 7 papers, 2 papers with code

Curvature-corrected learning dynamics in deep neural networks

no code implementations ICML 2020 Dongsung Huh

However, block-diagonal approximations of natural gradient, which are widely used in most second order methods (e. g. K-FAC), significantly distort the dynamics to follow highly divergent paths, destroying weight balance across layers.

Second-order methods

Discovering Symmetry Group Structures via Implicit Orthogonality Bias

no code implementations26 Feb 2024 Dongsung Huh

We introduce the HyperCube network, a novel approach for autonomously discovering symmetry group structures within data.

Inductive Bias

ISAAC Newton: Input-based Approximate Curvature for Newton's Method

1 code implementation1 May 2023 Felix Petersen, Tobias Sutter, Christian Borgelt, Dongsung Huh, Hilde Kuehne, Yuekai Sun, Oliver Deussen

We present ISAAC (Input-baSed ApproximAte Curvature), a novel method that conditions the gradient using selected second-order information and has an asymptotically vanishing computational overhead, assuming a batch size smaller than the number of neurons.

Second-order methods

The Missing Invariance Principle Found -- the Reciprocal Twin of Invariant Risk Minimization

1 code implementation29 May 2022 Dongsung Huh, Avinash Baidya

Further, we introduce a simplified, practical version of the MRI formulation called MRI-v1.

Few-Shot Multi-task Learning via Implicit regularization

no code implementations29 Sep 2021 Dongsung Huh

However, FSL problems proves to be significantly more challenging and require more compute expensive process to optimize.

Few-Shot Learning Multi-Task Learning

EXACT ANALYSIS OF CURVATURE CORRECTED LEARNING DYNAMICS IN DEEP LINEAR NETWORKS

no code implementations25 Sep 2019 Dongsung Huh

Deep neural networks exhibit complex learning dynamics due to the highly non-convex loss landscape, which causes slow convergence and vanishing gradient problems.

Second-order methods

Gradient Descent for Spiking Neural Networks

no code implementations NeurIPS 2018 Dongsung Huh, Terrence J. Sejnowski

Much of studies on neural computation are based on network models of static neurons that produce analog output, despite the fact that information processing in the brain is predominantly carried out by dynamic neurons that produce discrete pulses called spikes.

Cannot find the paper you are looking for? You can Submit a new open access paper.