Search Results for author: Benjamin K. Bergen

Found 18 papers, 9 papers with code

Revenge of the Fallen? Recurrent Models Match Transformers at Predicting Human Language Comprehension Metrics

no code implementations30 Apr 2024 James A. Michaelov, Catherine Arnett, Benjamin K. Bergen

Transformers have supplanted Recurrent Neural Networks as the dominant architecture for both natural language processing tasks and, despite criticisms of cognitive implausibility, for modelling the effect of predictability on online human language comprehension.

A Bit of a Problem: Measurement Disparities in Dataset Sizes Across Languages

1 code implementation1 Mar 2024 Catherine Arnett, Tyler A. Chang, Benjamin K. Bergen

We release a tool to obtain byte premiums for any two languages, enabling comparisons of dataset sizes across languages for more equitable multilingual model development and data practices.

When Is Multilinguality a Curse? Language Modeling for 250 High- and Low-Resource Languages

1 code implementation15 Nov 2023 Tyler A. Chang, Catherine Arnett, Zhuowen Tu, Benjamin K. Bergen

However, concrete evidence for the effects of multilinguality on language modeling performance in individual languages remains scarce.

Language Modelling

Structural Priming Demonstrates Abstract Grammatical Representations in Multilingual Language Models

no code implementations15 Nov 2023 James A. Michaelov, Catherine Arnett, Tyler A. Chang, Benjamin K. Bergen

We measure crosslingual structural priming in large language models, comparing model behavior to human experimental results from eight crosslingual experiments covering six languages, and four monolingual structural priming experiments in three non-English languages.

Sentence

Does GPT-4 pass the Turing test?

no code implementations31 Oct 2023 Cameron R. Jones, Benjamin K. Bergen

We evaluated GPT-4 in a public online Turing test.

Crosslingual Structural Priming and the Pre-Training Dynamics of Bilingual Language Models

no code implementations11 Oct 2023 Catherine Arnett, Tyler A. Chang, James A. Michaelov, Benjamin K. Bergen

Do multilingual language models share abstract grammatical representations across languages, and if so, when do these develop?

Language Modelling

Characterizing Learning Curves During Language Model Pre-Training: Learning, Forgetting, and Stability

1 code implementation29 Aug 2023 Tyler A. Chang, Zhuowen Tu, Benjamin K. Bergen

We quantify the final surprisal, within-run variability, age of acquisition, forgettability, and cross-run variability of learning curves for individual tokens in context.

Language Modelling

Emergent inabilities? Inverse scaling over the course of pretraining

no code implementations24 May 2023 James A. Michaelov, Benjamin K. Bergen

Does inverse scaling only occur as a function of model parameter size, or can it also occur over the course of training?

Language Modelling Math

Language Model Behavior: A Comprehensive Survey

1 code implementation20 Mar 2023 Tyler A. Chang, Benjamin K. Bergen

Transformer language models have received widespread public attention, yet their generated text is often surprising even to NLP researchers.

Language Modelling Large Language Model +1

Can Peanuts Fall in Love with Distributional Semantics?

no code implementations20 Jan 2023 James A. Michaelov, Seana Coulson, Benjamin K. Bergen

Context changes expectations about upcoming words - following a story involving an anthropomorphic peanut, comprehenders expect the sentence the peanut was in love more than the peanut was salted, as indexed by N400 amplitude (Nieuwland & van Berkum, 2006).

Sentence

Collateral facilitation in humans and language models

1 code implementation9 Nov 2022 James A. Michaelov, Benjamin K. Bergen

Are the predictions of humans and language models affected by similar things?

XLM-R

The Geometry of Multilingual Language Model Representations

1 code implementation22 May 2022 Tyler A. Chang, Zhuowen Tu, Benjamin K. Bergen

The subspace means differ along language-sensitive axes that are relatively stable throughout middle layers, and these axes encode information such as token vocabularies.

Cross-Lingual Transfer Transfer Learning +1

Word Acquisition in Neural Language Models

1 code implementation5 Oct 2021 Tyler A. Chang, Benjamin K. Bergen

We investigate how neural language models acquire individual words during training, extracting learning curves and ages of acquisition for over 600 words on the MacArthur-Bates Communicative Development Inventory (Fenson et al., 2007).

Language Acquisition

So Cloze yet so Far: N400 Amplitude is Better Predicted by Distributional Information than Human Predictability Judgements

no code implementations2 Sep 2021 James A. Michaelov, Seana Coulson, Benjamin K. Bergen

In this study, we investigate whether the linguistic predictions of computational language models or humans better reflect the way in which natural language stimuli modulate the amplitude of the N400.

Different kinds of cognitive plausibility: why are transformers better than RNNs at predicting N400 amplitude?

no code implementations20 Jul 2021 James A. Michaelov, Megan D. Bardolph, Seana Coulson, Benjamin K. Bergen

Despite being designed for performance rather than cognitive plausibility, transformer language models have been found to be better at predicting metrics used to assess human language comprehension than language models with other architectures, such as recurrent neural networks.

How well does surprisal explain N400 amplitude under different experimental conditions?

1 code implementation9 Oct 2020 James A. Michaelov, Benjamin K. Bergen

We investigate the extent to which word surprisal can be used to predict a neural measure of human language processing difficulty - the N400.

Cannot find the paper you are looking for? You can Submit a new open access paper.