no code implementations • LREC 2022 • Ariel Ekgren, Amaru Cuba Gyllensten, Evangelia Gogoulou, Alice Heiman, Severine Verlinden, Joey Öhman, Fredrik Carlsson, Magnus Sahlgren
We present GTP-SW3, a 3. 5 billion parameter autoregressive language model, trained on a newly created 100 GB Swedish corpus.
no code implementations • 2 Nov 2023 • Evangelia Gogoulou, Timothée Lesort, Magnus Boman, Joakim Nivre
The recent increase in data and model scale for language model pre-training has led to huge training costs.
no code implementations • 22 May 2023 • Ariel Ekgren, Amaru Cuba Gyllensten, Felix Stollenwerk, Joey Öhman, Tim Isbister, Evangelia Gogoulou, Fredrik Carlsson, Alice Heiman, Judit Casademont, Magnus Sahlgren
This paper details the process of developing the first native large generative language model for the Nordic languages, GPT-SW3.
no code implementations • 30 Mar 2023 • Joey Öhman, Severine Verlinden, Ariel Ekgren, Amaru Cuba Gyllensten, Tim Isbister, Evangelia Gogoulou, Fredrik Carlsson, Magnus Sahlgren
Pre-training Large Language Models (LLMs) require massive amounts of text data, and the performance of the LLMs typically correlates with the scale and quality of the datasets.
no code implementations • LREC 2022 • Evangelia Gogoulou, Ariel Ekgren, Tim Isbister, Magnus Sahlgren
Additionally, the results of evaluating the transferred models in source language tasks reveal that their performance in the source domain deteriorates after transfer.
no code implementations • EACL 2021 • Evangelia Gogoulou, Magnus Boman, Fehmi ben Abdesslem, Nils Hentati Isacsson, Viktor Kaldo, Magnus Sahlgren
We investigate the feasibility of applying standard text categorisation methods to patient text in order to predict treatment outcome in Internet-based cognitive behavioural therapy.
1 code implementation • ICLR 2021 • Fredrik Carlsson, Amaru Cuba Gyllensten, Evangelia Gogoulou, Erik Ylipää Hellqvist, Magnus Sahlgren
Extracting semantically useful natural language sentence representations from pre-trained deep neural networks such as Transformers remains a challenge.
no code implementations • SEMEVAL 2020 • Amaru Cuba Gyllensten, Evangelia Gogoulou, Ariel Ekgren, Magnus Sahlgren
We (Team Skurt) propose a simple method to detect lexical semantic change by clustering contextualized embeddings produced by XLM-R, using K-Means++.