Search Results for author: Zhenwei An

Found 4 papers, 4 papers with code

Harder Tasks Need More Experts: Dynamic Routing in MoE Models

1 code implementation12 Mar 2024 Quzhe Huang, Zhenwei An, Nan Zhuang, Mingxu Tao, Chen Zhang, Yang Jin, Kun Xu, Liwei Chen, Songfang Huang, Yansong Feng

In this paper, we introduce a novel dynamic expert selection framework for Mixture of Experts (MoE) models, aiming to enhance computational efficiency and model performance by adjusting the number of activated experts based on input difficulty.

Computational Efficiency

A Step Closer to Comprehensive Answers: Constrained Multi-Stage Question Decomposition with Large Language Models

1 code implementation13 Nov 2023 Hejing Cao, Zhenwei An, Jiazhan Feng, Kun Xu, Liwei Chen, Dongyan Zhao

While large language models exhibit remarkable performance in the Question Answering task, they are susceptible to hallucinations.

Question Answering

Lawyer LLaMA Technical Report

1 code implementation24 May 2023 Quzhe Huang, Mingxu Tao, Chen Zhang, Zhenwei An, Cong Jiang, Zhibin Chen, Zirui Wu, Yansong Feng

Specifically, we inject domain knowledge during the continual training stage and teach the model to learn professional skills using properly designed supervised fine-tuning tasks.

Hallucination Llama +1

Do Charge Prediction Models Learn Legal Theory?

1 code implementation31 Oct 2022 Zhenwei An, Quzhe Huang, Cong Jiang, Yansong Feng, Dongyan Zhao

The charge prediction task aims to predict the charge for a case given its fact description.

Cannot find the paper you are looking for? You can Submit a new open access paper.