1 code implementation • 28 Oct 2023 • Yangjun Wu, Kebin Fang, Dongxiang Zhang, Han Wang, Hao Zhang, Gang Chen
Structured dropout approaches, such as attention dropout and DropHead, have been investigated to regularize the multi-head attention mechanism in Transformers.
no code implementations • 9 Nov 2022 • Yangjun Wu, Kebin Fang, Yao Zhao, Hao Zhang, Lifeng Shi, Mengqi Zhang
To accomplish punctuation restoration, most existing methods focus on introducing extra information (e. g., part-of-speech) or addressing the class imbalance problem.
1 code implementation • 23 Mar 2022 • Yangjun Wu, Kebin Fang, Yao Zhao
To accomplish the punctuation restoration task, most existing approaches focused on leveraging extra information (e. g., part-of-speech tags) or addressing the class imbalance problem.