1 code implementation • 14 Nov 2023 • Pengyu Cheng, Yifan Yang, Jian Li, Yong Dai, Tianhao Hu, Peixin Cao, Nan Du
Human preference alignment is essential to improve the interaction quality of large language models (LLMs).