no code implementations • 6 Sep 2023 • Aobo Xia, Shuyu Lei, Yushu Yang, Xiang Guo, Hua Chai
This paper explores the instruction fine-tuning technique for speech-to-semantic tasks by introducing a unified end-to-end (E2E) framework that generates target text conditioned on a task-related prompt for audio data.
1 code implementation • 4 Feb 2021 • Han Liu, Caixia Yuan, Xiaojie Wang, Yushu Yang, Huixing Jiang, Zhongyuan Wang
We propose a novel task, Multi-Document Driven Dialogue (MD3), in which an agent can guess the target document that the user is interested in by leading a dialogue.
no code implementations • COLING 2020 • Keqing He, Shuyu Lei, Yushu Yang, Huixing Jiang, Zhongyuan Wang
Slot filling and intent detection are two major tasks for spoken language understanding.
1 code implementation • 1 Oct 2020 • Zipeng Xu, Fangxiang Feng, Xiaojie Wang, Yushu Yang, Huixing Jiang, Zhongyuan Wang
In this paper, we propose an Answer-Driven Visual State Estimator (ADVSE) to impose the effects of different answers on visual states.