no code implementations • ACL (WAT) 2021 • Raj Dabre, Abhisek Chakrabarty
The objective of the task was to explore the utility of multilingual approaches using a variety of in-domain and out-of-domain parallel and monolingual corpora.
no code implementations • AACL (WAT) 2020 • Raj Dabre, Abhisek Chakrabarty
In this paper we describe our team‘s (NICT-5) Neural Machine Translation (NMT) models whose translations were submitted to shared tasks of the 7th Workshop on Asian Translation.
no code implementations • COLING 2022 • Abhisek Chakrabarty, Raj Dabre, Chenchen Ding, Hideki Tanaka, Masao Utiyama, Eiichiro Sumita
In this paper we present FeatureBART, a linguistically motivated sequence-to-sequence monolingual pre-training strategy in which syntactic features such as lemma, part-of-speech and dependency labels are incorporated into the span prediction based pre-training framework (BART).
no code implementations • 15 Apr 2021 • Raj Dabre, Aizhan Imankulova, Masahiro Kaneko, Abhisek Chakrabarty
Parallel corpora are indispensable for training neural machine translation (NMT) models, and parallel corpora for most language pairs do not exist or are scarce.
no code implementations • COLING 2020 • Abhisek Chakrabarty, Raj Dabre, Chenchen Ding, Masao Utiyama, Eiichiro Sumita
In this study, linguistic knowledge at different levels are incorporated into the neural machine translation (NMT) framework to improve translation quality for language pairs with extremely limited data.
no code implementations • ACL 2017 • Abhisek Chakrabarty, P, Onkar Arun it, Utpal Garain
It is found that except Bengali, the proposed method outperforms Lemming and Morfette on the other languages.
no code implementations • LREC 2016 • Abhisek Chakrabarty, Akshay Chaturvedi, Utpal Garain
Given a word along with its contextual neighbours as input, the model is designed to produce the lemma of the concerned word as output.