no code implementations • 9 Mar 2024 • Yudong Yang, Rongfeng Su, Xiaokang Liu, Nan Yan, Lan Wang
In this model, the inherent acoustic characteristics of individuals related to the tongue motion details are encoded by using wav2vec 2. 0, while the ASR transcriptions related to the universality of tongue motions are encoded by using BERT.
1 code implementation • 20 Apr 2023 • Xiaokang Liu, Jianquan Li, Jingjing Mu, Min Yang, Ruifeng Xu, Benyou Wang
In this paper, we introduce novel K-center contrastive learning and adjustable decision boundary learning (CLAB) to improve the effectiveness of open intent classification.
1 code implementation • 12 Mar 2023 • Haofei Zhang, Mengqi Xue, Xiaokang Liu, KaiXuan Chen, Jie Song, Mingli Song
In this paper, we study a novel inference paradigm, termed as schema inference, that learns to deductively infer the explainable predictions by rebuilding the prior deep neural network (DNN) forwarding scheme, guided by the prevalent philosophical cognitive concept of schema.
1 code implementation • 2 Jul 2022 • Benyou Wang, Xiangbo Wu, Xiaokang Liu, Jianquan Li, Prayag Tiwari, Qianqian Xie
However, the humor aspect of natural language is relatively under-investigated, especially in the age of pre-trained language models.
1 code implementation • EMNLP 2020 • Jianquan Li, Xiaokang Liu, Honghong Zhao, Ruifeng Xu, Min Yang, Yaohong Jin
In this way, our model can learn from different teacher layers adaptively for various NLP tasks.
1 code implementation • 24 Sep 2020 • Xiaokang Liu, Haijun Song
The machine classifier exhibited 0. 99 precision on minerals, such as dolomite and pyrite.
no code implementations • 10 Mar 2020 • Xiaokang Liu, Shujie Ma, Kun Chen
We propose a nested reduced-rank regression (NRRR) approach in fitting regression model with multivariate functional responses and predictors, to achieve tailored dimension reduction and facilitate interpretation/visualization of the resulting functional model.
no code implementations • 16 Aug 2019 • Jianquan Li, Xiaokang Liu, Wenpeng Yin, Min Yang, Liqun Ma, Yaohong Jin
Multi-Task Learning (MTL) aims at boosting the overall performance of each individual task by leveraging useful information contained in multiple related tasks.
1 code implementation • 26 Jul 2018 • Gen Li, Xiaokang Liu, Kun Chen
Multi-view data have been routinely collected in various fields of science and engineering.
no code implementations • 24 Jan 2018 • Xuejing Yuan, Yuxuan Chen, Yue Zhao, Yunhui Long, Xiaokang Liu, Kai Chen, Shengzhi Zhang, Heqing Huang, Xiao-Feng Wang, Carl A. Gunter
For this purpose, we developed novel techniques that address a key technical challenge: integrating the commands into a song in a way that can be effectively recognized by ASR through the air, in the presence of background noise, while not being detected by a human listener.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1