1 code implementation • ACL 2022 • Fredrik Carlsson, Joey Öhman, Fangyu Liu, Severine Verlinden, Joakim Nivre, Magnus Sahlgren
We propose a resource-efficient method for converting a pre-trained CLM into this architecture, and demonstrate its potential on various experiments, including the novel task of contextualized word inclusion.
1 code implementation • LREC 2022 • Fredrik Carlsson, Philipp Eisen, Faton Rekathati, Magnus Sahlgren
The long-standing endeavor of relating the textual and the visual domain recently underwent a pivotal breakthrough, as OpenAI released CLIP.
Ranked #4 on Zero-shot Image Retrieval on XTD10
no code implementations • NoDaLiDa 2021 • Magnus Sahlgren, Fredrik Carlsson, Fredrik Olsson, Love Börjeson
When is it beneficial for a research community to organize a broader collaborative effort on a topic, and when should we instead promote individual efforts?
no code implementations • EMNLP (MRQA) 2021 • Fredrik Carlsson, Magnus Sahlgren, Fredrik Olsson, Amaru Cuba Gyllensten
This paper introduces a long-range multiple-choice Question Answering (QA) dataset, based on full-length fiction book texts.
no code implementations • LREC 2022 • Ariel Ekgren, Amaru Cuba Gyllensten, Evangelia Gogoulou, Alice Heiman, Severine Verlinden, Joey Öhman, Fredrik Carlsson, Magnus Sahlgren
We present GTP-SW3, a 3. 5 billion parameter autoregressive language model, trained on a newly created 100 GB Swedish corpus.
no code implementations • 22 May 2023 • Ariel Ekgren, Amaru Cuba Gyllensten, Felix Stollenwerk, Joey Öhman, Tim Isbister, Evangelia Gogoulou, Fredrik Carlsson, Alice Heiman, Judit Casademont, Magnus Sahlgren
This paper details the process of developing the first native large generative language model for the Nordic languages, GPT-SW3.
no code implementations • 30 Mar 2023 • Joey Öhman, Severine Verlinden, Ariel Ekgren, Amaru Cuba Gyllensten, Tim Isbister, Evangelia Gogoulou, Fredrik Carlsson, Magnus Sahlgren
Pre-training Large Language Models (LLMs) require massive amounts of text data, and the performance of the LLMs typically correlates with the scale and quality of the datasets.
1 code implementation • NoDaLiDa 2021 • Tim Isbister, Fredrik Carlsson, Magnus Sahlgren
We demonstrate empirically that a large English language model coupled with modern machine translation outperforms native language models in most Scandinavian languages.
no code implementations • 8 Feb 2021 • Magnus Sahlgren, Fredrik Carlsson
By contrast, we will argue that there are many different types of language use, meaning and understanding, and that (current) language models are build with the explicit purpose of acquiring and representing one type of structural understanding of language.
1 code implementation • ICLR 2021 • Fredrik Carlsson, Amaru Cuba Gyllensten, Evangelia Gogoulou, Erik Ylipää Hellqvist, Magnus Sahlgren
Extracting semantically useful natural language sentence representations from pre-trained deep neural networks such as Transformers remains a challenge.