Search Results for author: Jindřich Libovický

Found 34 papers, 11 papers with code

The LMU Munich System for the WMT 2021 Large-Scale Multilingual Machine Translation Shared Task

no code implementations WMT (EMNLP) 2021 Wen Lai, Jindřich Libovický, Alexander Fraser

This paper describes the submission of LMU Munich to the WMT 2021 multilingual machine translation task for small track #1, which studies translation between 6 languages (Croatian, Hungarian, Estonian, Serbian, Macedonian, English) in 30 directions.

Data Augmentation Knowledge Distillation +2

Why don’t people use character-level machine translation?

no code implementations Findings (ACL) 2022 Jindřich Libovický, Helmut Schmid, Alexander Fraser

We present a literature and empirical survey that critically assesses the state of the art in character-level modeling for machine translation (MT).

Machine Translation Translation

The LMU Munich Systems for the WMT21 Unsupervised and Very Low-Resource Translation Task

no code implementations WMT (EMNLP) 2021 Jindřich Libovický, Alexander Fraser

We present our submissions to the WMT21 shared task in Unsupervised and Very Low Resource machine translation between German and Upper Sorbian, German and Lower Sorbian, and Russian and Chuvash.

Machine Translation Translation

Charles Translator: A Machine Translation System between Ukrainian and Czech

no code implementations10 Apr 2024 Martin Popel, Lucie Poláková, Michal Novák, Jindřich Helcl, Jindřich Libovický, Pavel Straňák, Tomáš Krabač, Jaroslava Hlaváčová, Mariia Anisimova, Tereza Chlaňová

We present Charles Translator, a machine translation system between Ukrainian and Czech, developed as part of a society-wide effort to mitigate the impact of the Russian-Ukrainian war on individuals and society.

Translation Transliteration

Understanding Cross-Lingual Alignment -- A Survey

no code implementations9 Apr 2024 Katharina Hämmerl, Jindřich Libovický, Alexander Fraser

Cross-lingual alignment, the meaningful similarity of representations across languages in multilingual language models, has been an active field of research in recent years.

Decoder

How Gender Interacts with Political Values: A Case Study on Czech BERT Models

no code implementations20 Mar 2024 Adnan Al Ali, Jindřich Libovický

This case study focuses on the political biases of pre-trained encoders in Czech and compares them with a representative value survey.

CUNI Submission to MRL 2023 Shared Task on Multi-lingual Multi-task Information Retrieval

no code implementations25 Oct 2023 Jindřich Helcl, Jindřich Libovický

The goal of the shared task was to develop systems for named entity recognition and question answering in several under-represented languages.

Information Retrieval Machine Translation +5

A Dataset and Strong Baselines for Classification of Czech News Texts

1 code implementation20 Jul 2023 Hynek Kydlíček, Jindřich Libovický

Pre-trained models for Czech Natural Language Processing are often evaluated on purely linguistic tasks (POS tagging, parsing, NER) and relatively simple classification tasks such as sentiment classification or article classification from a single news source.

Classification NER +5

Is a Prestigious Job the same as a Prestigious Country? A Case Study on Multilingual Sentence Embeddings and European Countries

no code implementations23 May 2023 Jindřich Libovický

We study how multilingual sentence representations capture European countries and occupations and how this differs across European languages.

Sentence Sentence Embeddings

Probing the Role of Positional Information in Vision-Language Models

no code implementations Findings (NAACL) 2022 Philipp J. Rösch, Jindřich Libovický

Our results thus highlight an important issue of multimodal modeling: the mere presence of information detectable by a probing classifier is not a guarantee that the information is available in a cross-modal setup.

Contrastive Learning Image-text matching +5

CUNI Systems for the WMT22 Czech-Ukrainian Translation Task

no code implementations1 Dec 2022 Martin Popel, Jindřich Libovický, Jindřich Helcl

We present Charles University submissions to the WMT22 General Translation Shared Task on Czech-Ukrainian and Ukrainian-Czech machine translation.

Machine Translation Translation

Do Multilingual Language Models Capture Differing Moral Norms?

no code implementations18 Mar 2022 Katharina Hämmerl, Björn Deiseroth, Patrick Schramowski, Jindřich Libovický, Alexander Fraser, Kristian Kersting

Massively multilingual sentence representations are trained on large corpora of uncurated data, with a very imbalanced proportion of languages included in the training.

Sentence XLM-R

Improving Both Domain Robustness and Domain Adaptability in Machine Translation

1 code implementation COLING 2022 Wen Lai, Jindřich Libovický, Alexander Fraser

First, we want to reach domain robustness, i. e., we want to reach high quality on both domains seen in the training data and unseen domains.

Domain Adaptation Machine Translation +3

Why don't people use character-level machine translation?

no code implementations15 Oct 2021 Jindřich Libovický, Helmut Schmid, Alexander Fraser

We present a literature and empirical survey that critically assesses the state of the art in character-level modeling for machine translation (MT).

Machine Translation Translation

Neural String Edit Distance

1 code implementation spnlp (ACL) 2022 Jindřich Libovický, Alexander Fraser

We propose the neural string edit distance model for string-pair matching and string transduction based on learnable string edit distance.

Classification General Classification +1

Towards Reasonably-Sized Character-Level Transformer NMT by Finetuning Subword Systems

2 code implementations EMNLP 2020 Jindřich Libovický, Alexander Fraser

Applying the Transformer architecture on the character level usually requires very deep architectures that are difficult and slow to train.

Machine Translation NMT +2

Improving Fluency of Non-Autoregressive Machine Translation

no code implementations7 Apr 2020 Zdeněk Kasner, Jindřich Libovický, Jindřich Helcl

Non-autoregressive (nAR) models for machine translation (MT) manifest superior decoding speed when compared to autoregressive (AR) models, at the expense of impaired fluency of their outputs.

Machine Translation Translation

How Language-Neutral is Multilingual BERT?

1 code implementation8 Nov 2019 Jindřich Libovický, Rudolf Rosa, Alexander Fraser

Multilingual BERT (mBERT) provides sentence representations for 104 languages, which are useful for many multi-lingual tasks.

Retrieval Sentence +2

Probing Representations Learned by Multimodal Recurrent and Transformer Models

no code implementations29 Aug 2019 Jindřich Libovický, Pranava Madhyastha

In this paper, we present a meta-study assessing the representational quality of models where the training signal is obtained from different modalities, in particular, language modeling, image features prediction, and both textual and multimodal machine translation.

Image Retrieval Language Modelling +5

Neural Networks as Explicit Word-Based Rules

no code implementations10 Jul 2019 Jindřich Libovický

Filters of convolutional networks used in computer vision are often visualized as image patches that maximize the response of the filter.

General Classification Sentiment Analysis +1

End-to-End Non-Autoregressive Neural Machine Translation with Connectionist Temporal Classification

1 code implementation12 Nov 2018 Jindřich Libovický, Jindřich Helcl

Autoregressive decoding is the only part of sequence-to-sequence models that prevents them from massive parallelization at inference time.

Decoder General Classification +2

CUNI System for the WMT18 Multimodal Translation Task

no code implementations12 Nov 2018 Jindřich Helcl, Jindřich Libovický, Dušan Variš

For our submission, we acquired both textual and multimodal additional data.

Translation

Input Combination Strategies for Multi-Source Transformer Decoder

no code implementations12 Nov 2018 Jindřich Libovický, Jindřich Helcl, David Mareček

In multi-source sequence-to-sequence tasks, the attention mechanism can be modeled in several ways.

Decoder Translation

CUNI System for the WMT17 Multimodal Translation Task

no code implementations14 Jul 2017 Jindřich Helcl, Jindřich Libovický

For Task 1 (multimodal translation), our best scoring system is a purely textual neural translation of the source image caption to the target language.

Image Captioning Task 2 +1

Attention Strategies for Multi-Source Sequence-to-Sequence Learning

1 code implementation21 Apr 2017 Jindřich Libovický, Jindřich Helcl

Modeling attention in neural multi-source sequence-to-sequence learning remains a relatively unexplored area, despite its usefulness in tasks that incorporate multiple source languages or modalities.

Automatic Post-Editing Translation

CUNI System for WMT16 Automatic Post-Editing and Multimodal Translation Tasks

no code implementations WS 2016 Jindřich Libovický, Jindřich Helcl, Marek Tlustý, Pavel Pecina, Ondřej Bojar

Neural sequence to sequence learning recently became a very promising paradigm in machine translation, achieving competitive results with statistical phrase-based systems.

Automatic Post-Editing Multimodal Machine Translation +1

Cannot find the paper you are looking for? You can Submit a new open access paper.