Word Alignment

84 papers with code • 7 benchmarks • 4 datasets

Word Alignment is the task of finding the correspondence between source and target words in a pair of sentences that are translations of each other.

Source: Neural Network-based Word Alignment through Score Aggregation

Latest papers with no code

Learning Trajectory-Word Alignments for Video-Language Tasks

no code yet • ICCV 2023

To amend this, we propose a novel TW-BERT to learn Trajectory-Word alignment by a newly designed trajectory-to-word (T2W) attention for solving video-language tasks.

Word Alignment in the Era of Deep Learning: A Tutorial

no code yet • 30 Nov 2022

Jumping forward to the era of neural machine translation (NMT), we show how insights from word alignment inspired the attention mechanism fundamental to present-day NMT.

EntityCS: Improving Zero-Shot Cross-lingual Transfer with Entity-Centric Code Switching

no code yet • 22 Oct 2022

We use Wikidata and English Wikipedia to construct an entity-centric CS corpus by switching entities to their counterparts in other languages.

Extending Word-Level Quality Estimation for Post-Editing Assistance

no code yet • 23 Sep 2022

Based on extended word alignment, we further propose a novel task called refined word-level QE that outputs refined tags and word-level correspondences.

When do Contrastive Word Alignments Improve Many-to-many Neural Machine Translation?

no code yet • Findings (NAACL) 2022

Meanwhile, the contrastive objective can implicitly utilize automatically learned word alignment, which has not been explored in many-to-many NMT.

Graph Neural Networks for Multiparallel Word Alignment

no code yet • Findings (ACL) 2022

First, we create a multiparallel word alignment graph, joining all bilingual word alignment pairs in one graph.

Embedding-Enhanced GIZA++: Improving Low-Resource Word Alignment Using Embeddings

no code yet • ACL ARR January 2022

In the lowest-resource setting, we outperform GIZA++ by 8. 5, 10. 9, and 12 AER for Ro-En, De-En, and En-Fr, respectively.

When do Contrastive Word Alignments Improve Many-to-many Neural Machine Translation?

no code yet • ACL ARR January 2022

Meanwhile, the contrastive objective can implicitly utilize automatically learned word alignment, which has not been explored in many-to-many NMT.

Multi-Stage Framework with Refinement based Point Set Registration for Unsupervised Bi-Lingual Word Alignment

no code yet • ACL ARR November 2021

Cross-lingual alignment of word embeddings play an important role in knowledge transfer across languages, for improving machine translation and other multi-lingual applications.

ATOGAN:Adaptive Training Objective Generative Adversarial Network for Cross-lingual Word Alignment in Non-Isomorphic Embedding Spaces

no code yet • ACL ARR November 2021

Cross-lingual word alignment is a task for word translation from monolingual word embedding spaces of two languages.