no code implementations • CL (ACL) 2021 • Olga Majewska, Diana McCarthy, Jasper J. F. van den Bosch, Nikolaus Kriegeskorte, Ivan Vulić, Anna Korhonen
We demonstrate how the resultant data set can be used for fine-grained analyses and evaluation of representation learning models on the intrinsic tasks of semantic clustering and semantic similarity.
no code implementations • EMNLP 2020 • Qianchu Liu, Diana McCarthy, Anna Korhonen
One of the most powerful features of contextualized models is their dynamic embeddings for words in context, leading to state-of-the-art representations for context-aware lexical semantics.
no code implementations • 13 Dec 2021 • Qianchu Liu, Diana McCarthy, Anna Korhonen
Our findings demonstrate that models are usually not being tested for word-in-context semantics in the same way as humans are in these tasks, which helps us better understand the model-human gap.
1 code implementation • EMNLP 2021 • Qianchu Liu, Edoardo M. Ponti, Diana McCarthy, Ivan Vulić, Anna Korhonen
In order to address these gaps, we present AM2iCo (Adversarial and Multilingual Meaning in Context), a wide-coverage cross-lingual and multilingual evaluation set; it aims to faithfully assess the ability of state-of-the-art (SotA) representation models to understand the identity of word meaning in cross-lingual contexts for 14 language pairs.
1 code implementation • COLING 2020 • Olga Majewska, Ivan Vuli{\'c}, Diana McCarthy, Anna Korhonen
We present the first evaluation of the applicability of a spatial arrangement method (SpAM) to a typologically diverse language sample, and its potential to produce semantic evaluation resources to support multilingual NLP, with a focus on verb semantics.
no code implementations • LREC 2020 • Olga Majewska, Diana McCarthy, Jasper van den Bosch, Nikolaus Kriegeskorte, Ivan Vuli{\'c}, Anna Korhonen
We present a novel methodology for fast bottom-up creation of large-scale semantic similarity resources to support development and evaluation of NLP systems.
no code implementations • CONLL 2019 • Qianchu Liu, Diana McCarthy, Ivan Vuli{\'c}, Anna Korhonen
In this paper, we present a thorough investigation on methods that align pre-trained contextualized embeddings into shared cross-lingual context-aware embedding space, providing strong reference benchmarks for future context-aware crosslingual models.
no code implementations • SEMEVAL 2019 • Qianchu Liu, Diana McCarthy, Anna Korhonen
There is a growing awareness of the need to handle rare and unseen words in word representation modelling.
no code implementations • LREC 2014 • Marianna Apidianaki, Emilia Verzeni, Diana McCarthy
Paraphrases extracted from parallel corpora by the pivot method (Bannard and Callison-Burch, 2005) constitute a valuable resource for multilingual NLP applications.