Search Results for author: Yijia Zhang

Found 12 papers, 4 papers with code

BitDistiller: Unleashing the Potential of Sub-4-Bit LLMs via Self-Distillation

1 code implementation16 Feb 2024 Dayou Du, Yijia Zhang, Shijie Cao, Jiaqi Guo, Ting Cao, Xiaowen Chu, Ningyi Xu

The upscaling of Large Language Models (LLMs) has yielded impressive advances in natural language processing, yet it also poses significant deployment challenges.

Knowledge Distillation Quantization

AFPQ: Asymmetric Floating Point Quantization for LLMs

1 code implementation3 Nov 2023 Yijia Zhang, Sicheng Zhang, Shijie Cao, Dayou Du, Jianyu Wei, Ting Cao, Ningyi Xu

Large language models (LLMs) show great performance in various tasks, but face deployment challenges from limited memory capacity and bandwidth.

Quantization

A Transformer-Based Model With Self-Distillation for Multimodal Emotion Recognition in Conversations

1 code implementation31 Oct 2023 Hui Ma, Jian Wang, Hongfei Lin, Bo Zhang, Yijia Zhang, Bo Xu

Emotion recognition in conversations (ERC), the task of recognizing the emotion of each utterance in a conversation, is crucial for building empathetic machines.

Emotion Recognition in Conversation Multimodal Emotion Recognition

TKwinFormer: Top k Window Attention in Vision Transformers for Feature Matching

no code implementations29 Aug 2023 Yun Liao, Yide Di, Hao Zhou, Kaijun Zhu, Mingyu Lu, Yijia Zhang, Qing Duan, Junhui Liu

Local feature matching remains a challenging task, primarily due to difficulties in matching sparse keypoints and low-texture regions.

Adam Accumulation to Reduce Memory Footprints of both Activations and Gradients for Large-scale DNN Training

no code implementations31 May 2023 Yijia Zhang, Yibo Han, Shijie Cao, Guohao Dai, Youshan Miao, Ting Cao, Fan Yang, Ningyi Xu

We find that previous gradient accumulation reduces activation memory but fails to be compatible with gradient memory reduction due to a contradiction between preserving gradients and releasing gradients.

Integer or Floating Point? New Outlooks for Low-Bit Quantization on Large Language Models

no code implementations21 May 2023 Yijia Zhang, Lingran Zhao, Shijie Cao, WenQiang Wang, Ting Cao, Fan Yang, Mao Yang, Shanghang Zhang, Ningyi Xu

In this study, we conduct a comparative analysis of INT and FP quantization with the same bit-width, revealing that the optimal quantization format varies across different layers due to the complexity and diversity of tensor distribution.

Quantization

TC-GAT: Graph Attention Network for Temporal Causality Discovery

no code implementations21 Apr 2023 Xiaosong Yuan, Ke Chen, Wanli Zuo, Yijia Zhang

The present study explores the intricacies of causal relationship extraction, a vital component in the pursuit of causality knowledge.

Graph Attention

A Unified Review of Deep Learning for Automated Medical Coding

no code implementations8 Jan 2022 Shaoxiong Ji, Wei Sun, Xiaobo Li, Hang Dong, Ara Taalas, Yijia Zhang, Honghan Wu, Esa Pitkänen, Pekka Marttinen

Automated medical coding, an essential task for healthcare operation and delivery, makes unstructured data manageable by predicting medical codes from clinical documents.

Exploring Semi-supervised Variational Autoencoders for Biomedical Relation Extraction

no code implementations18 Jan 2019 Yijia Zhang, Zhiyong Lu

Experimental results show that our method effectively exploits the unlabeled data to improve the performance and reduce the dependence on labeled data.

Relation Relation Extraction

Cannot find the paper you are looking for? You can Submit a new open access paper.