Search Results for author: Hande Celikkanat

Found 5 papers, 2 papers with code

A Closer Look at Parameter Contributions When Training Neural Language and Translation Models

no code implementations COLING 2022 Raúl Vázquez, Hande Celikkanat, Vinit Ravishankar, Mathias Creutz, Jörg Tiedemann

We analyze the learning dynamics of neural language and translation models using Loss Change Allocation (LCA), an indicator that enables a fine-grained analysis of parameter updates when optimizing for the loss function.

Causal Language Modeling Language Modelling +3

Uncertainty-Aware Natural Language Inference with Stochastic Weight Averaging

1 code implementation10 Apr 2023 Aarne Talman, Hande Celikkanat, Sami Virpioja, Markus Heinonen, Jörg Tiedemann

This paper introduces Bayesian uncertainty modeling using Stochastic Weight Averaging-Gaussian (SWAG) in Natural Language Understanding (NLU) tasks.

Natural Language Inference Natural Language Understanding

On the differences between BERT and MT encoder spaces and how to address them in translation tasks

no code implementations ACL 2021 Ra{\'u}l V{\'a}zquez, Hande Celikkanat, Mathias Creutz, J{\"o}rg Tiedemann

Various studies show that pretrained language models such as BERT cannot straightforwardly replace encoders in neural machine translation despite their enormous success in other tasks.

Machine Translation NMT +1

Cannot find the paper you are looking for? You can Submit a new open access paper.