no code implementations • 5 Feb 2024 • Yongchang Hao, Yanshuai Cao, Lili Mou
The major reason is due to the quadratic memory and cubic time complexity to compute the inverse of the matrix.
1 code implementation • 5 Feb 2024 • Yongchang Hao, Yanshuai Cao, Lili Mou
Despite large neural networks demonstrating remarkable abilities to complete different tasks, they require excessive memory usage to store the optimization states for training.
1 code implementation • 17 Oct 2022 • Yongchang Hao, Yuxin Liu, Lili Mou
We additionally propose a simple modification to stabilize the RL training on non-parallel datasets with our induced reward function.
2 code implementations • 29 Sep 2022 • Yuqiao Wen, Yongchang Hao, Yanshuai Cao, Lili Mou
Open-domain dialogue systems aim to interact with humans through natural language texts in an open-ended fashion.
no code implementations • ACL 2022 • Wenxuan Wang, Wenxiang Jiao, Yongchang Hao, Xing Wang, Shuming Shi, Zhaopeng Tu, Michael Lyu
In this paper, we present a substantial step in better understanding the SOTA sequence-to-sequence (Seq2Seq) pretraining for neural machine translation~(NMT).
1 code implementation • NAACL 2021 • Yongchang Hao, Shilin He, Wenxiang Jiao, Zhaopeng Tu, Michael Lyu, Xing Wang
In addition, experimental results demonstrate that our Multi-Task NAT is complementary to knowledge distillation, the standard knowledge transfer method for NAT.