Morphological Inflection
37 papers with code • 0 benchmarks • 1 datasets
Morphological Inflection is the task of generating a target (inflected form) word from a source word (base form), given a morphological attribute, e.g. number, tense, and person etc. It is useful for alleviating data sparsity issues in translating morphologically rich languages. The transformation from a base form to an inflected form usually includes concatenating the base form with a prefix or a suffix and substituting some characters. For example, the inflected form of a Finnish stem eläkeikä (retirement age) is eläkeiittä when the case is abessive and the number is plural.
Source: Tackling Sequence to Sequence Mapping Problems with Neural Networks
Benchmarks
These leaderboards are used to track progress in Morphological Inflection
Latest papers
Falling Through the Gaps: Neural Architectures as Models of Morphological Rule Learning
We evaluate the Transformer as a model of morphological rule learning and compare it with Recurrent Neural Networks (RNN) on English, German, and Russian.
Minimal Supervision for Morphological Inflection
Neural models for the various flavours of morphological inflection tasks have proven to be extremely accurate given ample labeled data -- data that may be slow and costly to obtain.
On Biasing Transformer Attention Towards Monotonicity
Many sequence-to-sequence tasks in natural language processing are roughly monotonic in the alignment between source and target sequence, and previous work has facilitated or enforced learning of monotonic attention behavior via specialized attention functions or pretraining.
Interpretability for Morphological Inflection: from Character-level Predictions to Subword-level Rules
We apply our methodology to analyze the model{'}s decisions on three typologically-different languages and find that a) our pattern extraction method applied to cross-attention weights uncovers variation in form of inflection morphemes, b) pattern extraction from self-attention shows triggers for such variation, c) both types of patterns are closely aligned with grammar inflection classes and class assignment criteria, for all three languages.
Smoothing and Shrinking the Sparse Seq2Seq Search Space
Current sequence-to-sequence models are trained to minimize cross-entropy and use softmax to compute the locally normalized probabilities over target sequences.
SIGMORPHON 2020 Shared Task 0: Typologically Diverse Morphological Inflection
Systems were developed using data from 45 languages and just 5 language families, fine-tuned with data from an additional 45 languages and 10 language families (13 in total), and evaluated on all 90 languages.
Applying the Transformer to Character-level Transduction
The transformer has been shown to outperform recurrent neural network-based sequence-to-sequence models in various word-level NLP tasks.
CAMeL Tools: An Open Source Python Toolkit for Arabic Natural Language Processing
We present CAMeL Tools, a collection of open-source tools for Arabic natural language processing in Python.
Mind Your Inflections! Improving NLP for Non-Standard Englishes with Base-Inflection Encoding
Inflectional variation is a common feature of World Englishes such as Colloquial Singapore English and African American Vernacular English.
A Latent Morphology Model for Open-Vocabulary Neural Machine Translation
Translation into morphologically-rich languages challenges neural machine translation (NMT) models with extremely sparse vocabularies where atomic treatment of surface forms is unrealistic.