Search Results for author: Simran Kaur

Found 5 papers, 2 papers with code

Skill-Mix: a Flexible and Expandable Family of Evaluations for AI models

no code implementations26 Oct 2023 Dingli Yu, Simran Kaur, Arushi Gupta, Jonah Brown-Cohen, Anirudh Goyal, Sanjeev Arora

The paper develops a methodology for (a) designing and administering such an evaluation, and (b) automatic grading (plus spot-checking by humans) of the results using GPT-4 as well as the open LLaMA-2 70B model.

Disentangling the Mechanisms Behind Implicit Regularization in SGD

1 code implementation29 Nov 2022 Zachary Novack, Simran Kaur, Tanya Marwah, Saurabh Garg, Zachary C. Lipton

A number of competing hypotheses have been proposed to explain why small-batch Stochastic Gradient Descent (SGD)leads to improved generalization over the full-batch regime, with recent work crediting the implicit regularization of various quantities throughout training.

On the Maximum Hessian Eigenvalue and Generalization

no code implementations21 Jun 2022 Simran Kaur, Jeremy Cohen, Zachary C. Lipton

The mechanisms by which certain training interventions, such as increasing learning rates and applying batch normalization, improve the generalization of deep networks remains a mystery.

Gradient Descent on Neural Networks Typically Occurs at the Edge of Stability

1 code implementation ICLR 2021 Jeremy M. Cohen, Simran Kaur, Yuanzhi Li, J. Zico Kolter, Ameet Talwalkar

We empirically demonstrate that full-batch gradient descent on neural network training objectives typically operates in a regime we call the Edge of Stability.

Are Perceptually-Aligned Gradients a General Property of Robust Classifiers?

no code implementations18 Oct 2019 Simran Kaur, Jeremy Cohen, Zachary C. Lipton

For a standard convolutional neural network, optimizing over the input pixels to maximize the score of some target class will generally produce a grainy-looking version of the original image.

Adversarial Robustness

Cannot find the paper you are looking for? You can Submit a new open access paper.