Search Results for author: Madhur Panwar

Found 3 papers, 1 papers with code

Learning Syntax Without Planting Trees: Understanding When and Why Transformers Generalize Hierarchically

no code implementations25 Apr 2024 Kabir Ahuja, Vidhisha Balachandran, Madhur Panwar, Tianxing He, Noah A. Smith, Navin Goyal, Yulia Tsvetkov

Transformers trained on natural language data have been shown to learn its hierarchical structure and generalize to sentences with unseen syntactic structures without explicitly encoding any structural bias.

Inductive Bias Language Modelling

In-Context Learning through the Bayesian Prism

1 code implementation8 Jun 2023 Madhur Panwar, Kabir Ahuja, Navin Goyal

One of the main discoveries in this line of research has been that for several function classes, such as linear regression, transformers successfully generalize to new functions in the class.

Bayesian Inference In-Context Learning +4

Cannot find the paper you are looking for? You can Submit a new open access paper.