Search Results for author: Christian Herold

Found 20 papers, 1 papers with code

Recurrent Attention for the Transformer

no code implementations EMNLP (insights) 2021 Jan Rosendahl, Christian Herold, Frithjof Petrick, Hermann Ney

In this work, we conduct a comprehensive investigation on one of the centerpieces of modern machine translation systems: the encoder-decoder attention mechanism.

Machine Translation Translation

Detecting Various Types of Noise for Neural Machine Translation

no code implementations Findings (ACL) 2022 Christian Herold, Jan Rosendahl, Joris Vanvinckenroye, Hermann Ney

The filtering and/or selection of training data is one of the core aspects to be considered when building a strong machine translation system. In their influential work, Khayrallah and Koehn (2018) investigated the impact of different types of noise on the performance of machine translation systems. In the same year the WMT introduced a shared task on parallel corpus filtering, which went on to be repeated in the following years, and resulted in many different filtering approaches being proposed. In this work we aim to combine the recent achievements in data filtering with the original analysis of Khayrallah and Koehn (2018) and investigate whether state-of-the-art filtering systems are capable of removing all the suggested noise types. We observe that most of these types of noise can be detected with an accuracy of over 90% by modern filtering systems when operating in a well studied high resource setting. However, we also find that when confronted with more refined noise categories or when working with a less common language pair, the performance of the filtering systems is far from optimal, showing that there is still room for improvement in this area of research.

Machine Translation Translation

Improving Long Context Document-Level Machine Translation

no code implementations8 Jun 2023 Christian Herold, Hermann Ney

Document-level context for neural machine translation (NMT) is crucial to improve the translation consistency and cohesion, the translation of ambiguous inputs, as well as several other linguistic phenomena.

Document Level Machine Translation Machine Translation +3

On Search Strategies for Document-Level Neural Machine Translation

no code implementations8 Jun 2023 Christian Herold, Hermann Ney

On the other hand, in most works, the question on how to perform search with the trained model is scarcely discussed, sometimes not mentioned at all.

Machine Translation NMT +2

Does Joint Training Really Help Cascaded Speech Translation?

1 code implementation24 Oct 2022 Viet Anh Khoa Tran, David Thulke, Yingbo Gao, Christian Herold, Hermann Ney

Currently, in speech translation, the straightforward approach - cascading a recognition system with a translation system - delivers state-of-the-art results.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

Revisiting Checkpoint Averaging for Neural Machine Translation

no code implementations21 Oct 2022 Yingbo Gao, Christian Herold, Zijian Yang, Hermann Ney

Checkpoint averaging is a simple and effective method to boost the performance of converged neural machine translation models.

Machine Translation Translation

Diving Deep into Context-Aware Neural Machine Translation

no code implementations WMT (EMNLP) 2020 Jingjing Huo, Christian Herold, Yingbo Gao, Leonard Dahlmann, Shahram Khadivi, Hermann Ney

Context-aware neural machine translation (NMT) is a promising direction to improve the translation quality by making use of the additional context, e. g., document-level translation, or having meta-information.

Machine Translation NMT +1

Dynamic memory to alleviate catastrophic forgetting in continuous learning settings

no code implementations6 Jul 2020 Johannes Hofmanninger, Matthias Perkonigg, James A. Brink, Oleg Pianykh, Christian Herold, Georg Langs

In medical imaging, technical progress or changes in diagnostic procedures lead to a continuous change in image appearance.

Exploring Kernel Functions in the Softmax Layer for Contextual Word Classification

no code implementations EMNLP (IWSLT) 2019 Yingbo Gao, Christian Herold, Weiyue Wang, Hermann Ney

Prominently used in support vector machines and logistic regressions, kernel functions (kernels) can implicitly map data points into high dimensional spaces and make it easier to learn complex decision boundaries.

General Classification Language Modelling +2

The RWTH Aachen University Machine Translation Systems for WMT 2019

no code implementations WS 2019 Jan Rosendahl, Christian Herold, Yunsu Kim, Miguel Gra{\c{c}}a, Weiyue Wang, Parnia Bahar, Yingbo Gao, Hermann Ney

For the De-En task, none of the tested methods gave a significant improvement over last years winning system and we end up with the same performance, resulting in 39. 6{\%} BLEU on newstest2019.

Attribute Language Modelling +3

Improving Neural Language Models with Weight Norm Initialization and Regularization

no code implementations WS 2018 Christian Herold, Yingbo Gao, Hermann Ney

Embedding and projection matrices are commonly used in neural language models (NLM) as well as in other sequence processing networks that operate on large vocabularies.

Automatic Speech Recognition (ASR) Language Modelling +1

Cannot find the paper you are looking for? You can Submit a new open access paper.