no code implementations • 28 Mar 2024 • Kanishka Misra, Kyle Mahowald
Training on a corpus of human-scale in size (100M words), we iteratively trained transformer language models on systematically manipulated corpora and then evaluated their learning of a particular rare grammatical phenomenon: the English Article+Adjective+Numeral+Noun (AANN) construction (``a beautiful five days'').
no code implementations • 12 Jan 2024 • Kanishka Misra, Allyson Ettinger, Kyle Mahowald
Recent zero-shot evaluations have highlighted important limitations in the abilities of language models (LMs) to perform meaning extraction.
no code implementations • 3 Nov 2023 • Kanishka Misra, Najoung Kim
Exemplar based accounts are often considered to be in direct opposition to pure linguistic abstraction in explaining language learners' ability to generalize to novel expressions.
no code implementations • 6 Jun 2023 • Kanishka Misra, Cicero Nogueira dos santos, Siamak Shakeri
Despite readily memorizing world knowledge about entities, pre-trained language models (LMs) struggle to compose together two or more facts to perform multi-hop reasoning in question-answering tasks.
1 code implementation • 31 Jan 2023 • Freda Shi, Xinyun Chen, Kanishka Misra, Nathan Scales, David Dohan, Ed Chi, Nathanael Schärli, Denny Zhou
We use this benchmark to measure the distractibility of cutting-edge prompting techniques for large language models, and find that the model performance is dramatically decreased when irrelevant information is included.
no code implementations • 18 Dec 2022 • Koustuv Sinha, Jon Gauthier, Aaron Mueller, Kanishka Misra, Keren Fuentes, Roger Levy, Adina Williams
In this paper, we investigate the stability of language models' performance on targeted syntactic evaluations as we vary properties of the input context: the length of the context, the types of syntactic phenomena it contains, and whether or not there are violations of grammaticality.
1 code implementation • 5 Oct 2022 • Kanishka Misra, Julia Taylor Rayz, Allyson Ettinger
A characteristic feature of human semantic cognition is its ability to not only store and retrieve the properties of concepts observed through experience, but to also facilitate the inheritance of properties (can breathe) from superordinate concepts (animal) to their subordinates (dog) -- i. e. demonstrate property inheritance.
1 code implementation • 13 May 2022 • Kanishka Misra, Julia Taylor Rayz, Allyson Ettinger
To what extent can experience from language contribute to our conceptual knowledge?
1 code implementation • 24 Mar 2022 • Kanishka Misra
We present minicons, an open source library that provides a standard API for researchers interested in conducting behavioral and representational analyses of transformer-based language models (LMs).
no code implementations • 4 Nov 2021 • Kanishka Misra
My doctoral research focuses on understanding semantic knowledge in neural network models trained solely to predict natural language (referred to as language models, or LMs), by drawing on insights from the study of concepts and categories grounded in cognitive science.
1 code implementation • 6 May 2021 • Kanishka Misra, Allyson Ettinger, Julia Taylor Rayz
Building on research arguing for the possibility of conceptual and categorical knowledge acquisition through statistics contained in language, we evaluate predictive language models (LMs) -- informed solely by textual input -- on a prevalent phenomenon in cognitive science: typicality.
1 code implementation • 22 Apr 2021 • Kanishka Misra, Julia Taylor Rayz
Humans often communicate by using imprecise language, suggesting that fuzzy concepts with unclear boundaries are prevalent in language use.
no code implementations • 19 Jan 2021 • Qingyuan Hu, Yi Zhang, Kanishka Misra, Julia Rayz
Natural Language Inference (NLI) or Recognizing Textual Entailment (RTE) is the task of predicting the entailment relation between a pair of sentences (premise and hypothesis).
Natural Language Inference Natural Language Understanding +1
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Kanishka Misra, Allyson Ettinger, Julia Taylor Rayz
Models trained to estimate word probabilities in context have become ubiquitous in natural language processing.