Search Results for author: Surabhi Kumari

Found 5 papers, 0 papers with code

Domain Adaptation for NMT via Filtered Iterative Back-Translation

no code implementations EACL (AdaptNLP) 2021 Surabhi Kumari, Nikhil Jaiswal, Mayur Patidar, Manasi Patwardhan, Shirish Karande, Puneet Agarwal, Lovekesh Vig

In comparison, in this work, we observe that a simpler filtering approach based on a domain classifier, applied only to the pseudo-training data can consistently perform better, providing performance gains of 1. 40, 1. 82 and 0. 76 in terms of BLEU score for Medical, Law and IT in one direction, and 1. 28, 1. 60 and 1. 60 in the other direction in low resource scenario over competitive baselines.

Domain Adaptation Machine Translation +2

From Monolingual to Multilingual FAQ Assistant using Multilingual Co-training

no code implementations WS 2019 Mayur Patidar, Surabhi Kumari, Manasi Patwardhan, Kar, Shirish e, Puneet Agarwal, Lovekesh Vig, Gautam Shroff

We observe that the proposed approach provides consistent gains in the performance of BERT for multiple benchmark datasets (e. g. 1. 0{\%} gain on MLDocs, and 1. 2{\%} gain on XNLI over translate-train with BERT), while requiring a single model for multiple languages.

Cross-Lingual Transfer

Cannot find the paper you are looking for? You can Submit a new open access paper.