no code implementations • EMNLP (NLP-COVID19) 2020 • Zubair Afzal, Vikrant Yadav, Olga Fedorova, Vaishnavi Kandala, Janneke van de Loo, Saber A. Akhondi, Pascal Coupet, George Tsatsaronis
Ever since the COVID-19 pandemic broke out, the academic and scientific research community, as well as industry and governments around the world have joined forces in an unprecedented manner to fight the threat.
no code implementations • sdp (COLING) 2022 • Yury Kashnitsky, Drahomira Herrmannova, Anita de Waard, George Tsatsaronis, Catriona Catriona Fennell, Cyril Labbe
As a test set, the participants are provided with a 5x larger corpus of openly accessible human-written as well as generated papers from the same scientific domains of documents.
no code implementations • 24 Apr 2023 • Hosein Azarbonyad, Zubair Afzal, George Tsatsaronis
In this paper, we describe Topic Pages, an inventory of scientific concepts and information around them extracted from a large collection of scientific books and journals.
1 code implementation • ACL 2019 • Subhradeep Kayal, George Tsatsaronis
Distributed representation of words, or word embeddings, have motivated methods for calculating semantic representations of word sequences such as phrases, sentences and paragraphs.
1 code implementation • AKBC 2019 • Antonios Minas Krasakis, Evangelos Kanoulas, George Tsatsaronis
To tackle this problem, we propose a methodology for extending training datasets to arbitrarily big sizes and training complex, data-hungry models using weak supervision.
1 code implementation • COLING 2018 • Tirthankar Ghosal, Vignesh Edithal, Asif Ekbal, Pushpak Bhattacharyya, George Tsatsaronis, Srinivasa Satya Sameer Kumar Chivukula
The proposed method outperforms the existing state-of-the-art on a document-level novelty detection dataset by a margin of ∼5{\%} in terms of accuracy.
no code implementations • WS 2017 • Subhradeep Kayal, Zubair Afzal, George Tsatsaronis, Sophia Katrenko, Pascal Coupet, Marius Doornenbal, Michelle Gregory
In this paper we present a solution for tagging funding bodies and grants in scientific articles using a combination of trained sequential learning models, namely conditional random fields (CRF), hidden markov models (HMM) and maximum entropy models (MaxEnt), on a benchmark set created in-house.
no code implementations • 15 Jan 2014 • George Tsatsaronis, Iraklis Varlamis, Michalis Vazirgiannis
Without doubt, a measure of relatedness between text segments must take into account both the lexical and the semantic relatedness between words.