no code implementations • 31 Mar 2024 • Weihao Zeng, Dayuan Fu, Keqing He, Yejie Wang, Yukai Xu, Weiran Xu
Language models pre-trained on general text have achieved impressive results in diverse fields.
no code implementations • 2 Mar 2024 • Weihao Zeng, Keqing He, Yejie Wang, Dayuan Fu, Weiran Xu
Pre-trained language models have been successful in many scenarios.
no code implementations • 27 Feb 2024 • Pei Wang, Keqing He, Yejie Wang, Xiaoshuai Song, Yutao Mou, Jingang Wang, Yunsen Xian, Xunliang Cai, Weiran Xu
Out-of-domain (OOD) intent detection aims to examine whether the user's query falls outside the predefined domain of the system, which is crucial for the proper functioning of task-oriented dialogue (TOD) systems.
1 code implementation • 18 Feb 2024 • Dayuan Fu, Jianzhao Huang, Siyuan Lu, Guanting Dong, Yejie Wang, Keqing He, Weiran Xu
Addressing the discrepancies between predictions and actual outcomes often aids individuals in expanding their thought processes and engaging in reflection, thereby facilitating reasoning in the correct direction.
no code implementations • 17 Feb 2024 • Pei Wang, Yejie Wang, Muxi Diao, Keqing He, Guanting Dong, Weiran Xu
In this work, we focus on improving the confidence estimation of large language models.
no code implementations • 14 Feb 2024 • Yejie Wang, Keqing He, Guanting Dong, Pei Wang, Weihao Zeng, Muxi Diao, Yutao Mou, Mengdi Zhang, Jingang Wang, Xunliang Cai, Weiran Xu
It learns diverse instruction targets and combines a code evaluation objective to enhance its code generation ability.
1 code implementation • 17 Jun 2023 • Weihao Zeng, Keqing He, Yejie Wang, Chen Zeng, Jingang Wang, Yunsen Xian, Weiran Xu
Pre-trained language models based on general text enable huge success in the NLP scenario.