Search Results for author: Baohao Liao

Found 11 papers, 3 papers with code

Is It a Free Lunch for Removing Outliers during Pretraining?

no code implementations19 Feb 2024 Baohao Liao, Christof Monz

With the growing size of large language models, the role of quantization becomes increasingly significant.

Quantization

ApiQ: Finetuning of 2-Bit Quantized Large Language Model

no code implementations7 Feb 2024 Baohao Liao, Christof Monz

Memory-efficient finetuning of large language models (LLMs) has recently attracted huge attention with the increasing size of LLMs, primarily due to the constraints posed by GPU memory limitations and the comparable results of these methods with full finetuning.

Language Modelling Large Language Model +1

ITEm: Unsupervised Image-Text Embedding Learning for eCommerce

no code implementations22 Oct 2023 Baohao Liao, Michael Kozielski, Sanjika Hewavitharana, Jiangbo Yuan, Shahram Khadivi, Tomer Lancewicki

How to teach a model to learn embedding from different modalities without neglecting information from the less dominant modality is challenging.

Ask Language Model to Clean Your Noisy Translation Data

no code implementations20 Oct 2023 Quinten Bolding, Baohao Liao, Brandon James Denis, Jun Luo, Christof Monz

Lastly, experiments on C-MTNT showcased its effectiveness in evaluating the robustness of NMT models, highlighting the potential of advanced language models for data cleaning and emphasizing C-MTNT as a valuable resource.

Language Modelling Machine Translation +2

Make Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuning

1 code implementation NeurIPS 2023 Baohao Liao, Shaomu Tan, Christof Monz

One effective way to reduce the activation memory is to apply a reversible model, so the intermediate activations are not necessary to be cached and can be recomputed.

Image Classification Question Answering

Parameter-Efficient Fine-Tuning without Introducing New Latency

no code implementations26 May 2023 Baohao Liao, Yan Meng, Christof Monz

Parameter-efficient fine-tuning (PEFT) of pre-trained language models has recently demonstrated remarkable achievements, effectively matching the performance of full fine-tuning while utilizing significantly fewer trainable parameters, and consequently addressing the storage and communication constraints.

Federated Learning

Mask More and Mask Later: Efficient Pre-training of Masked Language Models by Disentangling the [MASK] Token

1 code implementation9 Nov 2022 Baohao Liao, David Thulke, Sanjika Hewavitharana, Hermann Ney, Christof Monz

We show: (1) [MASK]s can indeed be appended at a later layer, being disentangled from the word embedding; (2) The gathering of contextualized information from unmasked tokens can be conducted with a few layers.

Back-translation for Large-Scale Multilingual Machine Translation

1 code implementation WMT (EMNLP) 2021 Baohao Liao, Shahram Khadivi, Sanjika Hewavitharana

Surprisingly, the smaller size of vocabularies perform better, and the extensive monolingual English data offers a modest improvement.

Machine Translation Translation

Unifying Input and Output Smoothing in Neural Machine Translation

no code implementations COLING 2020 Yingbo Gao, Baohao Liao, Hermann Ney

Soft contextualized data augmentation is a recent method that replaces one-hot representation of words with soft posterior distributions of an external language model, smoothing the input of neural machine translation systems.

Data Augmentation Language Modelling +2

Multi-Agent Mutual Learning at Sentence-Level and Token-Level for Neural Machine Translation

no code implementations Findings of the Association for Computational Linguistics 2020 Baohao Liao, Yingbo Gao, Hermann Ney

Mutual learning, where multiple agents learn collaboratively and teach one another, has been shown to be an effective way to distill knowledge for image classification tasks.

Image Classification Machine Translation +2

Cannot find the paper you are looking for? You can Submit a new open access paper.