Search Results for author: Dana Angluin

Found 6 papers, 1 papers with code

Transformers as Transducers

no code implementations2 Apr 2024 Lena Strobl, Dana Angluin, David Chiang, Jonathan Rawski, Ashish Sabharwal

We study the sequence-to-sequence mapping capacity of transformers by relating them to finite transducers, and find that they can express surprisingly large classes of transductions.

Hard Attention POS

Transformers as Recognizers of Formal Languages: A Survey on Expressivity

no code implementations1 Nov 2023 Lena Strobl, William Merrill, Gail Weiss, David Chiang, Dana Angluin

As transformers have gained prominence in natural language processing, some researchers have investigated theoretically what problems they can and cannot solve, by treating problems as formal languages.

Masked Hard-Attention Transformers and Boolean RASP Recognize Exactly the Star-Free Languages

no code implementations21 Oct 2023 Dana Angluin, David Chiang, Andy Yang

We consider transformer encoders with hard attention (in which all attention is focused on exactly one position) and strict future masking (in which each position only attends to positions strictly to its left), and prove that the class of languages recognized by these networks is exactly the star-free languages.

Hard Attention Position

Formal Language Recognition by Hard Attention Transformers: Perspectives from Circuit Complexity

no code implementations13 Apr 2022 Yiding Hao, Dana Angluin, Robert Frank

This paper analyzes three formal models of Transformer encoders that differ in the form of their self-attention mechanism: unique hard attention (UHAT); generalized unique hard attention (GUHAT), which generalizes UHAT; and averaging hard attention (AHAT).

Hard Attention

Regular omega-Languages with an Informative Right Congruence

no code implementations10 Sep 2018 Dana Angluin, Dana Fisman

The right congruence of a regular omega-language is not informative enough; many regular omega-languages have a trivial right congruence, and in general it is not always possible to define an omega-automaton recognizing a given language that is isomorphic to the rightcon automaton.

Cannot find the paper you are looking for? You can Submit a new open access paper.