Search Results for author: Wanshun Chen

Found 1 papers, 1 papers with code

On Diversified Preferences of Large Language Model Alignment

1 code implementation12 Dec 2023 Dun Zeng, Yong Dai, Pengyu Cheng, Longyue Wang, Tianhao Hu, Wanshun Chen, Nan Du, Zenglin Xu

Our analysis reveals a correlation between the calibration performance of reward models (RMs) and the alignment performance of LLMs.

Language Modelling Large Language Model

Cannot find the paper you are looking for? You can Submit a new open access paper.