no code implementations • EAMT 2022 • Gabriele Sarti, Arianna Bisazza
Neural machine translation (NMT) systems are nowadays essential components of professional translation workflows.
1 code implementation • NAACL (CMCL) 2021 • Gabriele Sarti, Dominique Brunato, Felice Dell’Orletta
We then show the effectiveness of linguistic features when explicitly leveraged by a regression model for predicting sentence complexity and compare its results with the ones obtained by a fine-tuned neural language model.
no code implementations • 30 Apr 2024 • Javier Ferrando, Gabriele Sarti, Arianna Bisazza, Marta R. Costa-jussà
The rapid progress of research aimed at interpreting the inner workings of advanced language models has highlighted a need for contextualizing the insights gained from years of work in this area.
no code implementations • 5 Oct 2023 • Anna Langedijk, Hosein Mohebbi, Gabriele Sarti, Willem Zuidema, Jaap Jumelet
In recent years, many interpretability methods have been proposed to help interpret the internal states of Transformer-models, at different levels of precision and complexity.
2 code implementations • 2 Oct 2023 • Gabriele Sarti, Grzegorz Chrupała, Malvina Nissim, Arianna Bisazza
Establishing whether language models can use contextual information in a human-plausible way is important to ensure their trustworthiness in real-world settings.
1 code implementation • 1 Sep 2023 • Daniel Scalena, Gabriele Sarti, Malvina Nissim, Elisabetta Fersini
Due to language models' propensity to generate toxic or hateful responses, several techniques were developed to align model generations with users' preferences.
no code implementations • 26 May 2023 • Gabriele Sarti, Phu Mon Htut, Xing Niu, Benjamin Hsu, Anna Currey, Georgiana Dinu, Maria Nadejde
Attribute-controlled translation (ACT) is a subtask of machine translation that involves controlling stylistic or linguistic attributes (like formality and gender) of translation outputs.
1 code implementation • 28 Feb 2023 • Lukas Edman, Gabriele Sarti, Antonio Toral, Gertjan van Noord, Arianna Bisazza
Pretrained character-level and byte-level language models have been shown to be competitive with popular subword models across a range of Natural Language Processing (NLP) tasks.
2 code implementations • 27 Feb 2023 • Gabriele Sarti, Nils Feldhus, Ludwig Sickert, Oskar van der Wal, Malvina Nissim, Arianna Bisazza
Past work in natural language processing interpretability focused mainly on popular classification tasks while largely overlooking generation settings, partly due to a lack of dedicated tools.
1 code implementation • 24 May 2022 • Gabriele Sarti, Arianna Bisazza, Ana Guerberof Arenas, Antonio Toral
We publicly release the complete dataset including all collected behavioral data, to foster new research on the translation capabilities of NMT systems for typologically diverse languages.
3 code implementations • 7 Mar 2022 • Gabriele Sarti, Malvina Nissim
The T5 model and its unified text-to-text paradigm contributed in advancing the state-of-the-art for many natural language processing tasks.
1 code implementation • 19 Aug 2021 • Federico Bianchi, Giuseppe Attanasio, Raphael Pisoni, Silvia Terragni, Gabriele Sarti, Sri Lakshmi
CLIP (Contrastive Language-Image Pre-training) is a very recent multi-modal model that jointly learns representations of images and texts.
1 code implementation • NAACL (TeachingNLP) 2021 • Lucio Messina, Lucia Busso, Claudia Roberta Combei, Ludovica Pannitto, Alessio Miaschi, Gabriele Sarti, Malvina Nissim
We describe and make available the game-based material developed for a laboratory run at several Italian science festivals to popularize NLP among young students.
1 code implementation • NAACL (TeachingNLP) 2021 • Ludovica Pannitto, Lucia Busso, Claudia Roberta Combei, Lucio Messina, Alessio Miaschi, Gabriele Sarti, Malvina Nissim
To raise awareness, curiosity, and longer-term interest in young people, we have developed an interactive workshop designed to illustrate the basic principles of NLP and computational linguistics to high school Italian students aged between 13 and 18 years.
1 code implementation • 17 Dec 2020 • Jinen Setpal, Gabriele Sarti
We introduce ArchiMeDe, a multimodal neural network-based architecture used to solve the DANKMEMES meme detections subtask at the 2020 EVALITA campaign.
1 code implementation • 10 Nov 2020 • Gabriele Sarti
This work describes a self-supervised data augmentation approach used to improve learning models' performances when only a moderate amount of labeled data is available.
1 code implementation • 25 Aug 2020 • Ginevra Carbone, Gabriele Sarti
We first test the effectiveness of our approach in a low-resource setting for Italian, evaluating the conditioning for both topic models and gold annotations.