no code implementations • 1 Dec 2023 • Longfeng Nie, Yuntian Chen, Dongxiao Zhang, Xinyue Liu, Wentian Yuan
Specifically, the temporal and spatial characteristics of remote sensing data of the satellite Himawari-8 are extracted and fused by recurrent neural network (RNN) and convolution operation, respectively.
no code implementations • 4 Nov 2023 • Hang Yin, Yao Su, Xinyue Liu, Thomas Hartvigsen, Yanhua Li, Xiangnan Kong
We refer to such brain networks as multi-state, and this mixture can help us understand human behavior.
no code implementations • 13 Jan 2021 • Hang Yin, Xinyue Liu, Xiangnan Kong
Existing works mainly focus on unimodal distributions, where it is usually assumed that the observed activities aregenerated from asingleGaussian distribution (i. e., one graph). However, this assumption is too strong for many real-worldapplications.
no code implementations • COLING 2020 • Mingda Li, Xinyue Liu, Weitong Ruan, Luca Soldaini, Wael Hamza, Chengwei Su
The comparison shows that our model could recover the transcription by integrating the fragmented information among hypotheses and identifying the frequent error patterns of the ASR module, and even rewrite the query for a better understanding, which reveals the characteristic of multi-task learning of broadcasting knowledge.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +6
no code implementations • Findings of the Association for Computational Linguistics 2020 • Luoxin Chen, Xinyue Liu, Weitong Ruan, Jianhua Lu
Adversarial training (AT) has shown strong regularization effects on deep learning algorithms by introducing small input perturbations to improve model robustness.
Ranked #3 on Chunking on CoNLL 2000 (using extra training data)
no code implementations • ACL 2020 • Luoxin Chen, Weitong Ruan, Xinyue Liu, Jianhua Lu
Virtual adversarial training (VAT) is a powerful technique to improve model robustness in both supervised and semi-supervised settings.
Ranked #7 on Chunking on CoNLL 2000
no code implementations • 11 Jan 2020 • Mingda Li, Weitong Ruan, Xinyue Liu, Luca Soldaini, Wael Hamza, Chengwei Su
The NLU module usually uses the first best interpretation of a given speech in downstream tasks such as domain and intent classification.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +5
1 code implementation • 1 May 2019 • Thanh Tran, Xinyue Liu, Kyumin Lee, Xiangnan Kong
Personalized recommendation algorithms learn a user's preference for an item by measuring a distance/similarity between them.
no code implementations • 22 Aug 2018 • Xinyue Liu, Xiangnan Kong, Lei Liu, Kuorong Chiang
To address these issues, we study the problem of syntax-aware sequence generation with GANs, in which a collection of real sequences and a set of pre-defined grammatical rules are given to both discriminator and generator.