Search Results for author: Young Jin Kim

Found 18 papers, 9 papers with code

Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation

1 code implementation16 Jan 2024 Haoran Xu, Amr Sharaf, Yunmo Chen, Weiting Tan, Lingfeng Shen, Benjamin Van Durme, Kenton Murray, Young Jin Kim

However, even the top-performing 13B LLM-based translation models, like ALMA, does not match the performance of state-of-the-art conventional encoder-decoder translation models or larger-scale LLMs such as GPT-4.

Machine Translation Translation

PEMA: An Offsite-Tunable Plug-in External Memory Adaptation for Language Models

1 code implementation14 Nov 2023 HyunJin Kim, Young Jin Kim, JinYeong Bak

PEMA integrates with context representations from test data during inference to perform downstream tasks.

Machine Translation Sentence +1

Mixture of Quantized Experts (MoQE): Complementary Effect of Low-bit Quantization and Robustness

no code implementations3 Oct 2023 Young Jin Kim, Raffy Fahim, Hany Hassan Awadalla

In our comprehensive analysis, we show that MoE models with 2-bit expert weights can deliver better model performance than the dense model trained on the same dataset.

Machine Translation Quantization

A Paradigm Shift in Machine Translation: Boosting Translation Performance of Large Language Models

1 code implementation20 Sep 2023 Haoran Xu, Young Jin Kim, Amr Sharaf, Hany Hassan Awadalla

In this study, we propose a novel fine-tuning approach for LLMs that is specifically designed for the translation task, eliminating the need for the abundant parallel data that traditional translation models usually depend on.

Language Modelling Machine Translation +1

Task-Based MoE for Multitask Multilingual Machine Translation

no code implementations30 Aug 2023 Hai Pham, Young Jin Kim, Subhabrata Mukherjee, David P. Woodruff, Barnabas Poczos, Hany Hassan Awadalla

Mixture-of-experts (MoE) architecture has been proven a powerful method for diverse tasks in training deep models in many applications.

Machine Translation Translation

FineQuant: Unlocking Efficiency with Fine-Grained Weight-Only Quantization for LLMs

no code implementations16 Aug 2023 Young Jin Kim, Rawn Henry, Raffy Fahim, Hany Hassan Awadalla

Large Language Models (LLMs) have achieved state-of-the-art performance across various language tasks but pose challenges for practical deployment due to their substantial memory requirements.

Quantization

How Good Are GPT Models at Machine Translation? A Comprehensive Evaluation

1 code implementation18 Feb 2023 Amr Hendy, Mohamed Abdelrehim, Amr Sharaf, Vikas Raunak, Mohamed Gabr, Hitokazu Matsushita, Young Jin Kim, Mohamed Afify, Hany Hassan Awadalla

In this paper, we present a comprehensive evaluation of GPT models for machine translation, covering various aspects such as quality of different GPT models in comparison with state-of-the-art research and commercial systems, effect of prompting strategies, robustness towards domain shifts and document-level translation.

Machine Translation Text Generation +1

Who Says Elephants Can't Run: Bringing Large Scale MoE Models into Cloud Scale Production

no code implementations18 Nov 2022 Young Jin Kim, Rawn Henry, Raffy Fahim, Hany Hassan Awadalla

Mixture of Experts (MoE) models with conditional execution of sparsely activated layers have enabled training models with a much larger number of parameters.

Machine Translation

Gating Dropout: Communication-efficient Regularization for Sparsely Activated Transformers

no code implementations28 May 2022 Rui Liu, Young Jin Kim, Alexandre Muzio, Hany Hassan Awadalla

Sparsely activated transformers, such as Mixture of Experts (MoE), have received great interest due to their outrageous scaling capability which enables dramatical increases in model size without significant increases in computational cost.

Machine Translation

Taming Sparsely Activated Transformer with Stochastic Experts

1 code implementation ICLR 2022 Simiao Zuo, Xiaodong Liu, Jian Jiao, Young Jin Kim, Hany Hassan, Ruofei Zhang, Tuo Zhao, Jianfeng Gao

While most on-going research focuses on improving SAMs models by exploring methods of routing inputs to experts, our analysis reveals that such research might not lead to the solution we expect, i. e., the commonly-used routing methods based on gating mechanisms do not work better than randomly routing inputs to experts.

Machine Translation Translation

Scalable and Efficient MoE Training for Multitask Multilingual Models

1 code implementation22 Sep 2021 Young Jin Kim, Ammar Ahmad Awan, Alexandre Muzio, Andres Felipe Cruz Salinas, Liyang Lu, Amr Hendy, Samyam Rajbhandari, Yuxiong He, Hany Hassan Awadalla

By combining the efficient system and training methods, we are able to significantly scale up large multitask multilingual models for language generation which results in a great improvement in model accuracy.

Machine Translation Text Generation

FastFormers: Highly Efficient Transformer Models for Natural Language Understanding

2 code implementations EMNLP (sustainlp) 2020 Young Jin Kim, Hany Hassan Awadalla

In this paper, we present FastFormers, a set of recipes to achieve efficient inference-time performance for Transformer-based models on various NLU tasks.

Knowledge Distillation Natural Language Understanding

From Research to Production and Back: Ludicrously Fast Neural Machine Translation

no code implementations WS 2019 Young Jin Kim, Marcin Junczys-Dowmunt, Hany Hassan, Alham Fikri Aji, Kenneth Heafield, Roman Grundkiewicz, Nikolay Bogoychev

Taking our dominating submissions to the previous edition of the shared task as a starting point, we develop improved teacher-student training via multi-agent dual-learning and noisy backward-forward translation for Transformer-based student models.

C++ code Machine Translation +1

Automated cardiovascular magnetic resonance image analysis with fully convolutional networks

1 code implementation25 Oct 2017 Wenjia Bai, Matthew Sinclair, Giacomo Tarroni, Ozan Oktay, Martin Rajchl, Ghislain Vaillant, Aaron M. Lee, Nay Aung, Elena Lukaschuk, Mihir M. Sanghvi, Filip Zemrak, Kenneth Fung, Jose Miguel Paiva, Valentina Carapella, Young Jin Kim, Hideaki Suzuki, Bernhard Kainz, Paul M. Matthews, Steffen E. Petersen, Stefan K. Piechnik, Stefan Neubauer, Ben Glocker, Daniel Rueckert

By combining FCN with a large-scale annotated dataset, the proposed automated method achieves a high performance on par with human experts in segmenting the LV and RV on short-axis CMR images and the left atrium (LA) and right atrium (RA) on long-axis CMR images.

Cannot find the paper you are looking for? You can Submit a new open access paper.