Search Results for author: Chong Meng

Found 3 papers, 0 papers with code

Improving the Robustness of Large Language Models via Consistency Alignment

no code implementations21 Mar 2024 Yukun Zhao, Lingyong Yan, Weiwei Sun, Guoliang Xing, Shuaiqiang Wang, Chong Meng, Zhicong Cheng, Zhaochun Ren, Dawei Yin

The training process is accomplished by self-rewards inferred from the trained model at the first stage without referring to external human preference resources.

Instruction Following Response Generation

Knowing What LLMs DO NOT Know: A Simple Yet Effective Self-Detection Method

no code implementations27 Oct 2023 Yukun Zhao, Lingyong Yan, Weiwei Sun, Guoliang Xing, Chong Meng, Shuaiqiang Wang, Zhicong Cheng, Zhaochun Ren, Dawei Yin

In this paper, we propose a novel self-detection method to detect which questions that a LLM does not know that are prone to generate nonfactual results.

Cannot find the paper you are looking for? You can Submit a new open access paper.