Search Results for author: Like Hui

Found 5 papers, 0 papers with code

ReLU soothes the NTK condition number and accelerates optimization for wide neural networks

no code implementations15 May 2023 Chaoyue Liu, Like Hui

Comparing with linear neural networks, we show that a ReLU activated wide neural network at random initialization has a larger angle separation for similar data in the feature space of model gradient, and has a smaller condition number for NTK.

Cut your Losses with Squentropy

no code implementations8 Feb 2023 Like Hui, Mikhail Belkin, Stephen Wright

We provide an extensive set of experiments on multi-class classification problems showing that the squentropy loss outperforms both the pure cross entropy and rescaled square losses in terms of the classification accuracy.

Classification Multi-class Classification

Limitations of Neural Collapse for Understanding Generalization in Deep Learning

no code implementations17 Feb 2022 Like Hui, Mikhail Belkin, Preetum Nakkiran

We refine the Neural Collapse conjecture into two separate conjectures: collapse on the train set (an optimization property) and collapse on the test distribution (a generalization property).

Representation Learning

Evaluation of Neural Architectures Trained with Square Loss vs Cross-Entropy in Classification Tasks

no code implementations ICLR 2021 Like Hui, Mikhail Belkin

We explore several major neural architectures and a range of standard benchmark datasets for NLP, automatic speech recognition (ASR) and computer vision tasks to show that these architectures, with the same hyper-parameter settings as reported in the literature, perform comparably or better when trained with the square loss, even after equalizing computational resources.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

Cannot find the paper you are looking for? You can Submit a new open access paper.