Search Results for author: Kanishka Misra

Found 14 papers, 7 papers with code

Language Models Learn Rare Phenomena from Less Rare Phenomena: The Case of the Missing AANNs

no code implementations28 Mar 2024 Kanishka Misra, Kyle Mahowald

Training on a corpus of human-scale in size (100M words), we iteratively trained transformer language models on systematically manipulated corpora and then evaluated their learning of a particular rare grammatical phenomenon: the English Article+Adjective+Numeral+Noun (AANN) construction (``a beautiful five days'').

counterfactual Memorization

Experimental Contexts Can Facilitate Robust Semantic Property Inference in Language Models, but Inconsistently

no code implementations12 Jan 2024 Kanishka Misra, Allyson Ettinger, Kyle Mahowald

Recent zero-shot evaluations have highlighted important limitations in the abilities of language models (LMs) to perform meaning extraction.

Novel Concepts

Abstraction via exemplars? A representational case study on lexical category inference in BERT

no code implementations3 Nov 2023 Kanishka Misra, Najoung Kim

Exemplar based accounts are often considered to be in direct opposition to pure linguistic abstraction in explaining language learners' ability to generalize to novel expressions.

Triggering Multi-Hop Reasoning for Question Answering in Language Models using Soft Prompts and Random Walks

no code implementations6 Jun 2023 Kanishka Misra, Cicero Nogueira dos santos, Siamak Shakeri

Despite readily memorizing world knowledge about entities, pre-trained language models (LMs) struggle to compose together two or more facts to perform multi-hop reasoning in question-answering tasks.

Knowledge Graphs Question Answering +1

Large Language Models Can Be Easily Distracted by Irrelevant Context

1 code implementation31 Jan 2023 Freda Shi, Xinyun Chen, Kanishka Misra, Nathan Scales, David Dohan, Ed Chi, Nathanael Schärli, Denny Zhou

We use this benchmark to measure the distractibility of cutting-edge prompting techniques for large language models, and find that the model performance is dramatically decreased when irrelevant information is included.

Arithmetic Reasoning Language Modelling +1

Language model acceptability judgements are not always robust to context

no code implementations18 Dec 2022 Koustuv Sinha, Jon Gauthier, Aaron Mueller, Kanishka Misra, Keren Fuentes, Roger Levy, Adina Williams

In this paper, we investigate the stability of language models' performance on targeted syntactic evaluations as we vary properties of the input context: the length of the context, the types of syntactic phenomena it contains, and whether or not there are violations of grammaticality.

In-Context Learning Language Modelling +1

COMPS: Conceptual Minimal Pair Sentences for testing Robust Property Knowledge and its Inheritance in Pre-trained Language Models

1 code implementation5 Oct 2022 Kanishka Misra, Julia Taylor Rayz, Allyson Ettinger

A characteristic feature of human semantic cognition is its ability to not only store and retrieve the properties of concepts observed through experience, but to also facilitate the inheritance of properties (can breathe) from superordinate concepts (animal) to their subordinates (dog) -- i. e. demonstrate property inheritance.

Attribute

A Property Induction Framework for Neural Language Models

1 code implementation13 May 2022 Kanishka Misra, Julia Taylor Rayz, Allyson Ettinger

To what extent can experience from language contribute to our conceptual knowledge?

minicons: Enabling Flexible Behavioral and Representational Analyses of Transformer Language Models

1 code implementation24 Mar 2022 Kanishka Misra

We present minicons, an open source library that provides a standard API for researchers interested in conducting behavioral and representational analyses of transformer-based language models (LMs).

Benchmarking Sentence

On Semantic Cognition, Inductive Generalization, and Language Models

no code implementations4 Nov 2021 Kanishka Misra

My doctoral research focuses on understanding semantic knowledge in neural network models trained solely to predict natural language (referred to as language models, or LMs), by drawing on insights from the study of concepts and categories grounded in cognitive science.

Do language models learn typicality judgments from text?

1 code implementation6 May 2021 Kanishka Misra, Allyson Ettinger, Julia Taylor Rayz

Building on research arguing for the possibility of conceptual and categorical knowledge acquisition through statistics contained in language, we evaluate predictive language models (LMs) -- informed solely by textual input -- on a prevalent phenomenon in cognitive science: typicality.

Finding Fuzziness in Neural Network Models of Language Processing

1 code implementation22 Apr 2021 Kanishka Misra, Julia Taylor Rayz

Humans often communicate by using imprecise language, suggesting that fuzzy concepts with unclear boundaries are prevalent in language use.

Natural Language Inference

Exploring Lexical Irregularities in Hypothesis-Only Models of Natural Language Inference

no code implementations19 Jan 2021 Qingyuan Hu, Yi Zhang, Kanishka Misra, Julia Rayz

Natural Language Inference (NLI) or Recognizing Textual Entailment (RTE) is the task of predicting the entailment relation between a pair of sentences (premise and hypothesis).

Natural Language Inference Natural Language Understanding +1

Cannot find the paper you are looking for? You can Submit a new open access paper.