Search Results for author: Allyson Ettinger

Found 31 papers, 15 papers with code

“No, They Did Not”: Dialogue Response Dynamics in Pre-trained Language Models

no code implementations COLING 2022 Sanghee J. Kim, Lang Yu, Allyson Ettinger

A critical component of competence in language is being able to identify relevant components of an utterance and reply appropriately.

When Hindsight is Not 20/20: Testing Limits on Reflective Thinking in Large Language Models

no code implementations14 Apr 2024 Yanhong Li, Chenghao Yang, Allyson Ettinger

In this paper, we set out to clarify these capabilities under a more stringent evaluation setting in which we disallow any kind of external feedback.

Experimental Contexts Can Facilitate Robust Semantic Property Inference in Language Models, but Inconsistently

no code implementations12 Jan 2024 Kanishka Misra, Allyson Ettinger, Kyle Mahowald

Recent zero-shot evaluations have highlighted important limitations in the abilities of language models (LMs) to perform meaning extraction.

Novel Concepts

The Generative AI Paradox: "What It Can Create, It May Not Understand"

no code implementations31 Oct 2023 Peter West, Ximing Lu, Nouha Dziri, Faeze Brahman, Linjie Li, Jena D. Hwang, Liwei Jiang, Jillian Fisher, Abhilasha Ravichander, Khyathi Chandu, Benjamin Newman, Pang Wei Koh, Allyson Ettinger, Yejin Choi

Specifically, we propose and test the Generative AI Paradox hypothesis: generative models, having been trained directly to reproduce expert-like outputs, acquire generative capabilities that are not contingent upon -- and can therefore exceed -- their ability to understand those same types of outputs.

"You Are An Expert Linguistic Annotator": Limits of LLMs as Analyzers of Abstract Meaning Representation

no code implementations26 Oct 2023 Allyson Ettinger, Jena D. Hwang, Valentina Pyatkin, Chandra Bhagavatula, Yejin Choi

We compare models' analysis of this semantic structure across two settings: 1) direct production of AMR parses based on zero- and few-shot prompts, and 2) indirect partial reconstruction of AMR via metalinguistic natural language queries (e. g., "Identify the primary event of this sentence, and the predicate corresponding to that event.").

Natural Language Queries Sentence

Can You Follow Me? Testing Situational Understanding in ChatGPT

1 code implementation24 Oct 2023 Chenghao Yang, Allyson Ettinger

Understanding sentence meanings and updating information states appropriately across time -- what we call "situational understanding" (SU) -- is a critical ability for human-like AI agents.

Chatbot Sentence

Faith and Fate: Limits of Transformers on Compositionality

1 code implementation NeurIPS 2023 Nouha Dziri, Ximing Lu, Melanie Sclar, Xiang Lorraine Li, Liwei Jiang, Bill Yuchen Lin, Peter West, Chandra Bhagavatula, Ronan Le Bras, Jena D. Hwang, Soumya Sanyal, Sean Welleck, Xiang Ren, Allyson Ettinger, Zaid Harchaoui, Yejin Choi

We formulate compositional tasks as computation graphs to systematically quantify the level of complexity, and break down reasoning steps into intermediate sub-procedures.

Counterfactual reasoning: Testing language models' understanding of hypothetical scenarios

1 code implementation26 May 2023 Jiaxuan Li, Lang Yu, Allyson Ettinger

Current pre-trained language models have enabled remarkable improvements in downstream tasks, but it remains difficult to distinguish effects of statistical correlation from more systematic logical reasoning grounded on the understanding of real world.

counterfactual Counterfactual Reasoning +2

Counterfactual reasoning: Do language models need world knowledge for causal understanding?

1 code implementation6 Dec 2022 Jiaxuan Li, Lang Yu, Allyson Ettinger

Current pre-trained language models have enabled remarkable improvements in downstream tasks, but it remains difficult to distinguish effects of statistical correlation from more systematic logical reasoning grounded on understanding of the real world.

counterfactual Counterfactual Reasoning +2

"No, they did not": Dialogue response dynamics in pre-trained language models

1 code implementation5 Oct 2022 Sanghee J. Kim, Lang Yu, Allyson Ettinger

A critical component of competence in language is being able to identify relevant components of an utterance and reply appropriately.

COMPS: Conceptual Minimal Pair Sentences for testing Robust Property Knowledge and its Inheritance in Pre-trained Language Models

1 code implementation5 Oct 2022 Kanishka Misra, Julia Taylor Rayz, Allyson Ettinger

A characteristic feature of human semantic cognition is its ability to not only store and retrieve the properties of concepts observed through experience, but to also facilitate the inheritance of properties (can breathe) from superordinate concepts (animal) to their subordinates (dog) -- i. e. demonstrate property inheritance.

Attribute

A Property Induction Framework for Neural Language Models

1 code implementation13 May 2022 Kanishka Misra, Julia Taylor Rayz, Allyson Ettinger

To what extent can experience from language contribute to our conceptual knowledge?

Variation and generality in encoding of syntactic anomaly information in sentence embeddings

no code implementations EMNLP (BlackboxNLP) 2021 Qinxuan Wu, Allyson Ettinger

While sentence anomalies have been applied periodically for testing in NLP, we have yet to establish a picture of the precise status of anomaly information in representations from NLP models.

Anomaly Detection Sentence +1

Pragmatic competence of pre-trained language models through the lens of discourse connectives

no code implementations CoNLL (EMNLP) 2021 Lalchand Pandia, Yan Cong, Allyson Ettinger

We focus on testing models' ability to use pragmatic cues to predict discourse connectives, models' ability to understand implicatures relating to connectives, and the extent to which models show humanlike preferences regarding temporal dynamics of connectives.

Implicatures

Sorting through the noise: Testing robustness of information processing in pre-trained language models

no code implementations EMNLP 2021 Lalchand Pandia, Allyson Ettinger

Pre-trained LMs have shown impressive performance on downstream NLP tasks, but we have yet to establish a clear understanding of their sophistication when it comes to processing, retaining, and applying information presented in their input.

Semantic Similarity Semantic Textual Similarity

On the Interplay Between Fine-tuning and Composition in Transformers

2 code implementations Findings (ACL) 2021 Lang Yu, Allyson Ettinger

Here we investigate the impact of fine-tuning on the capacity of contextualized embeddings to capture phrase meaning information beyond lexical content.

Sentiment Analysis Sentiment Classification

Do language models learn typicality judgments from text?

1 code implementation6 May 2021 Kanishka Misra, Allyson Ettinger, Julia Taylor Rayz

Building on research arguing for the possibility of conceptual and categorical knowledge acquisition through statistics contained in language, we evaluate predictive language models (LMs) -- informed solely by textual input -- on a prevalent phenomenon in cognitive science: typicality.

Adding Recurrence to Pretrained Transformers

no code implementations1 Jan 2021 Davis Yoshida, Allyson Ettinger, Kevin Gimpel

Fine-tuning a pretrained transformer for a downstream task has become a standard method in NLP in the last few years.

Language Modelling

Assessing Phrasal Representation and Composition in Transformers

1 code implementation EMNLP 2020 Lang Yu, Allyson Ettinger

Deep transformer models have pushed performance on NLP tasks to new limits, suggesting sophisticated treatment of complex linguistic inputs, such as phrases.

Adding Recurrence to Pretrained Transformers for Improved Efficiency and Context Size

no code implementations16 Aug 2020 Davis Yoshida, Allyson Ettinger, Kevin Gimpel

Fine-tuning a pretrained transformer for a downstream task has become a standard method in NLP in the last few years.

Language Modelling

Spying on your neighbors: Fine-grained probing of contextual embeddings for information about surrounding words

no code implementations ACL 2020 Josef Klafka, Allyson Ettinger

Although models using contextual word embeddings have achieved state-of-the-art results on a host of NLP tasks, little is known about exactly what information these embeddings encode about the context words that they are understood to reflect.

Word Embeddings

What BERT is not: Lessons from a new suite of psycholinguistic diagnostics for language models

2 code implementations TACL 2020 Allyson Ettinger

Pre-training by language modeling has become a popular and successful approach to NLP tasks, but we have yet to understand exactly what linguistic capacities these pre-training processes confer upon models.

Language Modelling Negation

Assessing Composition in Sentence Vector Representations

1 code implementation COLING 2018 Allyson Ettinger, Ahmed Elgohary, Colin Phillips, Philip Resnik

We describe the details of the method and generation system, and then present results of experiments applying our method to probe for compositional information in embeddings from a number of existing sentence composition models.

Sentence

Towards Linguistically Generalizable NLP Systems: A Workshop and Shared Task

no code implementations WS 2017 Allyson Ettinger, Sudha Rao, Hal Daumé III, Emily M. Bender

This paper presents a summary of the first Workshop on Building Linguistically Generalizable Natural Language Processing Systems, and the associated Build It Break It, The Language Edition shared task.

Cannot find the paper you are looking for? You can Submit a new open access paper.