Search Results for author: Terra Blevins

Found 20 papers, 3 papers with code

MYTE: Morphology-Driven Byte Encoding for Better and Fairer Multilingual Language Modeling

no code implementations15 Mar 2024 Tomasz Limisiewicz, Terra Blevins, Hila Gonen, Orevaoghene Ahia, Luke Zettlemoyer

A major consideration in multilingual language modeling is how to best represent languages with diverse vocabularies and scripts.

Language Modelling

Comparing Hallucination Detection Metrics for Multilingual Generation

no code implementations16 Feb 2024 Haoqiang Kang, Terra Blevins, Luke Zettlemoyer

While many automatic hallucination detection techniques have been proposed for English texts, their effectiveness in multilingual contexts remains unexplored.

Hallucination Natural Language Inference +1

Breaking the Curse of Multilinguality with Cross-lingual Expert Language Models

no code implementations19 Jan 2024 Terra Blevins, Tomasz Limisiewicz, Suchin Gururangan, Margaret Li, Hila Gonen, Noah A. Smith, Luke Zettlemoyer

Despite their popularity in non-English NLP, multilingual language models often underperform monolingual ones due to inter-language competition for model parameters.

Detecting Pretraining Data from Large Language Models

no code implementations25 Oct 2023 Weijia Shi, Anirudh Ajith, Mengzhou Xia, Yangsibo Huang, Daogao Liu, Terra Blevins, Danqi Chen, Luke Zettlemoyer

Min-K% Prob can be applied without any knowledge about the pretraining corpus or any additional training, departing from previous detection methods that require training a reference model on data that is similar to the pretraining data.

Machine Unlearning

Embedding structure matters: Comparing methods to adapt multilingual vocabularies to new languages

1 code implementation9 Sep 2023 C. M. Downey, Terra Blevins, Nora Goldfine, Shane Steinert-Threlkeld

Pre-trained multilingual language models underpin a large portion of modern NLP tools outside of English.

Translate to Disambiguate: Zero-shot Multilingual Word Sense Disambiguation with Pretrained Language Models

no code implementations26 Apr 2023 Haoqiang Kang, Terra Blevins, Luke Zettlemoyer

To better understand this contrast, we present a new study investigating how well PLMs capture cross-lingual word sense with Contextual Word-Level Translation (C-WLT), an extension of word-level translation that prompts the model to translate a given word in context.

Translation Word Sense Disambiguation

Demystifying Prompts in Language Models via Perplexity Estimation

no code implementations8 Dec 2022 Hila Gonen, Srini Iyer, Terra Blevins, Noah A. Smith, Luke Zettlemoyer

Language models can be prompted to perform a wide variety of zero- and few-shot learning problems.

Few-Shot Learning

Prompting Language Models for Linguistic Structure

no code implementations15 Nov 2022 Terra Blevins, Hila Gonen, Luke Zettlemoyer

Although pretrained language models (PLMs) can be prompted to perform a wide range of language tasks, it remains an open question how much this ability comes from generalizable linguistic understanding versus surface-level lexical patterns.

Chunking In-Context Learning +8

Analyzing the Mono- and Cross-Lingual Pretraining Dynamics of Multilingual Language Models

no code implementations24 May 2022 Terra Blevins, Hila Gonen, Luke Zettlemoyer

The emergent cross-lingual transfer seen in multilingual pretrained models has sparked significant interest in studying their behavior.

Cross-Lingual Transfer XLM-R

Language Contamination Helps Explain the Cross-lingual Capabilities of English Pretrained Models

no code implementations17 Apr 2022 Terra Blevins, Luke Zettlemoyer

English pretrained language models, which make up the backbone of many modern NLP systems, require huge amounts of unlabeled training data.

Cross-Lingual Transfer

FEWS: Large-Scale, Low-Shot Word Sense Disambiguation with the Dictionary

no code implementations EACL 2021 Terra Blevins, Mandar Joshi, Luke Zettlemoyer

Current models for Word Sense Disambiguation (WSD) struggle to disambiguate rare senses, despite reaching human performance on global WSD metrics.

Transfer Learning Word Sense Disambiguation

Moving Down the Long Tail of Word Sense Disambiguation with Gloss Informed Bi-encoders

no code implementations ACL 2020 Terra Blevins, Luke Zettlemoyer

A major obstacle in Word Sense Disambiguation (WSD) is that word senses are not uniformly distributed, causing existing models to generally perform poorly on senses that are either rare or unseen during training.

Word Sense Disambiguation

Moving Down the Long Tail of Word Sense Disambiguation with Gloss-Informed Biencoders

1 code implementation6 May 2020 Terra Blevins, Luke Zettlemoyer

A major obstacle in Word Sense Disambiguation (WSD) is that word senses are not uniformly distributed, causing existing models to generally perform poorly on senses that are either rare or unseen during training.

Word Sense Disambiguation

Better Character Language Modeling Through Morphology

no code implementations ACL 2019 Terra Blevins, Luke Zettlemoyer

We incorporate morphological supervision into character language models (CLMs) via multitasking and show that this addition improves bits-per-character (BPC) performance across 24 languages, even when the morphology data and language modeling data are disjoint.

Language Modelling

Deep RNNs Encode Soft Hierarchical Syntax

no code implementations ACL 2018 Terra Blevins, Omer Levy, Luke Zettlemoyer

We present a set of experiments to demonstrate that deep recurrent neural networks (RNNs) learn internal representations that capture soft hierarchical notions of syntax from highly varied supervision.

Dependency Parsing Language Modelling +3

Automatically Processing Tweets from Gang-Involved Youth: Towards Detecting Loss and Aggression

no code implementations COLING 2016 Terra Blevins, Robert Kwiatkowski, Jamie MacBeth, Kathleen McKeown, Desmond Patton, Owen Rambow

Violence is a serious problems for cities like Chicago and has been exacerbated by the use of social media by gang-involved youths for taunting rival gangs.

Cannot find the paper you are looking for? You can Submit a new open access paper.