Search Results for author: Ahmet Üstün

Found 19 papers, 9 papers with code

Unsupervised Translation of German–Lower Sorbian: Exploring Training and Novel Transfer Methods on a Low-Resource Language

no code implementations WMT (EMNLP) 2021 Lukas Edman, Ahmet Üstün, Antonio Toral, Gertjan van Noord

This paper describes the methods behind the systems submitted by the University of Groningen for the WMT 2021 Unsupervised Machine Translation task for German–Lower Sorbian (DE–DSB): a high-resource language to a low-resource one.

Translation Unsupervised Machine Translation

Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs

no code implementations22 Feb 2024 Arash Ahmadian, Chris Cremer, Matthias Gallé, Marzieh Fadaee, Julia Kreutzer, Olivier Pietquin, Ahmet Üstün, Sara Hooker

AI alignment in the shape of Reinforcement Learning from Human Feedback (RLHF) is increasingly treated as a crucial ingredient for high performance large language models.

Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction Tuning

1 code implementation11 Sep 2023 Ted Zadouri, Ahmet Üstün, Arash Ahmadian, Beyza Ermiş, Acyr Locatelli, Sara Hooker

The Mixture of Experts (MoE) is a widely known neural architecture where an ensemble of specialized sub-models optimizes overall performance with a constant computational cost.

When Less is More: Investigating Data Pruning for Pretraining LLMs at Scale

no code implementations8 Sep 2023 Max Marion, Ahmet Üstün, Luiza Pozzobon, Alex Wang, Marzieh Fadaee, Sara Hooker

In this work, we take a wider view and explore scalable estimates of data quality that can be used to systematically measure the quality of pretraining data.

Memorization

When does Parameter-Efficient Transfer Learning Work for Machine Translation?

1 code implementation23 May 2022 Ahmet Üstün, Asa Cooper Stickland

We find that using PEFTs with a larger pre-trained model outperforms full fine-tuning with a smaller model, and for smaller training data sizes, PEFTs outperform full fine-tuning for the same pre-trained model.

Machine Translation Transfer Learning +1

Multilingual Unsupervised Neural Machine Translation with Denoising Adapters

no code implementations EMNLP 2021 Ahmet Üstün, Alexandre Bérard, Laurent Besacier, Matthias Gallé

We consider the problem of multilingual unsupervised machine translation, translating to and from languages that only have monolingual data by using auxiliary parallel language pairs.

Denoising Translation +1

Unsupervised Translation of German--Lower Sorbian: Exploring Training and Novel Transfer Methods on a Low-Resource Language

1 code implementation24 Sep 2021 Lukas Edman, Ahmet Üstün, Antonio Toral, Gertjan van Noord

Lastly, we experiment with the order in which offline and online back-translation are used to train an unsupervised system, finding that using online back-translation first works better for DE$\rightarrow$DSB by 2. 76 BLEU.

Translation Unsupervised Machine Translation

On the Difficulty of Translating Free-Order Case-Marking Languages

2 code implementations13 Jul 2021 Arianna Bisazza, Ahmet Üstün, Stephan Sportel

Identifying factors that make certain languages harder to model than others is essential to reach language equality in future Natural Language Processing technologies.

Machine Translation NMT +1

On the Effectiveness of Dataset Embeddings in Mono-lingual,Multi-lingual and Zero-shot Conditions

no code implementations EACL (AdaptNLP) 2021 Rob van der Goot, Ahmet Üstün, Barbara Plank

However, it remains unclear in which situations these dataset embeddings are most effective, because they are used in a large variety of settings, languages and tasks.

Dependency Parsing Lemmatization +1

FiSSA at SemEval-2020 Task 9: Fine-tuned For Feelings

1 code implementation SEMEVAL 2020 Bertelt Braaksma, Richard Scholtens, Stan van Suijlekom, Remy Wang, Ahmet Üstün

In this paper, we present our approach for sentiment classification on Spanish-English code-mixed social media data in the SemEval-2020 Task 9.

Classification General Classification +3

UDapter: Language Adaptation for Truly Universal Dependency Parsing

1 code implementation EMNLP 2020 Ahmet Üstün, Arianna Bisazza, Gosse Bouma, Gertjan van Noord

The resulting parser, UDapter, outperforms strong monolingual and multilingual baselines on the majority of both high-resource and low-resource (zero-shot) languages, showing the success of the proposed adaptation approach.

Dependency Parsing Transfer Learning

A Trie-Structured Bayesian Model for Unsupervised Morphological Segmentation

no code implementations24 Apr 2017 Murathan Kurfali, Ahmet Üstün, Burcu Can

Our results show that using different information sources such as neural word embeddings and letter successor variety as prior information improves morphological segmentation in a Bayesian model.

Segmentation Word Embeddings

Turkish PoS Tagging by Reducing Sparsity with Morpheme Tags in Small Datasets

no code implementations9 Mar 2017 Burcu Can, Ahmet Üstün, Murathan Kurfali

We learn inflectional and derivational morpheme tags in Turkish by using conditional random fields (CRF) and we employ the morpheme tags in part-of-speech (PoS) tagging by using hidden Markov models (HMMs) to mitigate sparsity.

Part-Of-Speech Tagging POS +1

Cannot find the paper you are looking for? You can Submit a new open access paper.