Search Results for author: En-Shiun Annie Lee

Found 7 papers, 1 papers with code

Unlocking Parameter-Efficient Fine-Tuning for Low-Resource Language Translation

no code implementations5 Apr 2024 Tong Su, Xin Peng, Sarubi Thillainathan, David Guzmán, Surangika Ranathunga, En-Shiun Annie Lee

Parameter-efficient fine-tuning (PEFT) methods are increasingly vital in adapting large-scale pre-trained language models for diverse tasks, offering a balance between adaptability and computational efficiency.

Computational Efficiency Machine Translation +2

Enhancing Hokkien Dual Translation by Exploring and Standardizing of Four Writing Systems

no code implementations18 Mar 2024 Bo-Han Lu, Yi-Hsuan Lin, En-Shiun Annie Lee, Richard Tzong-Han Tsai

We employ a pre-trained LLaMA2-7B model specialized in Traditional Mandarin Chinese to leverage the orthographic similarities between Taiwanese Hokkien Han and Traditional Mandarin Chinese.

Machine Translation Translation

Leveraging Auxiliary Domain Parallel Data in Intermediate Task Fine-tuning for Low-resource Translation

1 code implementation2 Jun 2023 Shravan Nayak, Surangika Ranathunga, Sarubi Thillainathan, Rikki Hung, Anthony Rinaldi, Yining Wang, Jonah Mackey, Andrew Ho, En-Shiun Annie Lee

In this paper, we show that intermediate-task fine-tuning (ITFT) of PMSS models is extremely beneficial for domain-specific NMT, especially when target domain data is limited/unavailable and the considered languages are missing or under-represented in the PMSS model.

NMT

Neural Machine Translation for Low-Resource Languages: A Survey

no code implementations29 Jun 2021 Surangika Ranathunga, En-Shiun Annie Lee, Marjana Prifti Skenduli, Ravi Shekhar, Mehreen Alam, Rishemjit Kaur

Neural Machine Translation (NMT) has seen a tremendous spurt of growth in less than ten years, and has already entered a mature phase.

Machine Translation NMT +1

Unsupervised Transfer Learning via BERT Neuron Selection

no code implementations10 Dec 2019 Mehrdad Valipour, En-Shiun Annie Lee, Jaime R. Jamacaro, Carolina Bessega

To determine whether there are task-specific neurons that can be exploited for unsupervised transfer learning, we introduce a method for selecting the most important neurons to solve a specific classification task.

Natural Language Inference Sentence +3

Cannot find the paper you are looking for? You can Submit a new open access paper.