Search Results for author: Du-Seong Chang

Found 6 papers, 5 papers with code

PSYDIAL: Personality-based Synthetic Dialogue Generation using Large Language Models

1 code implementation1 Apr 2024 Ji-Eun Han, Jun-Seok Koh, Hyeon-Tae Seo, Du-Seong Chang, Kyung-Ah Sohn

Experimental results indicate that while pre-trained models and those fine-tuned with a chit-chat dataset struggle to generate responses reflecting personality, models trained with PSYDIAL show significant improvements.

Dialogue Generation

ExeGPT: Constraint-Aware Resource Scheduling for LLM Inference

no code implementations15 Mar 2024 Hyungjun Oh, Kihong Kim, JaeMin Kim, Sungkyun Kim, Junyeol Lee, Du-Seong Chang, Jiwon Seo

This paper presents ExeGPT, a distributed system designed for constraint-aware LLM inference.

Scheduling

NASH: A Simple Unified Framework of Structured Pruning for Accelerating Encoder-Decoder Language Models

1 code implementation16 Oct 2023 Jongwoo Ko, Seungjoon Park, Yujin Kim, Sumyeong Ahn, Du-Seong Chang, Euijai Ahn, Se-Young Yun

Structured pruning methods have proven effective in reducing the model size and accelerating inference speed in various network architectures such as Transformers.

Decoder

Revisiting Intermediate Layer Distillation for Compressing Language Models: An Overfitting Perspective

1 code implementation3 Feb 2023 Jongwoo Ko, Seungjoon Park, Minchan Jeong, Sukjin Hong, Euijai Ahn, Du-Seong Chang, Se-Young Yun

Knowledge distillation (KD) is a highly promising method for mitigating the computational problems of pre-trained language models (PLMs).

Knowledge Distillation

Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders

1 code implementation20 Nov 2022 Minsoo Kim, Sihwa Lee, Sukjin Hong, Du-Seong Chang, Jungwook Choi

In particular, KD has been employed in quantization-aware training (QAT) of Transformer encoders like BERT to improve the accuracy of the student model with the reduced-precision weight parameters.

Knowledge Distillation Model Compression +1

Cannot find the paper you are looking for? You can Submit a new open access paper.