Search Results for author: Haibo Zhang

Found 22 papers, 7 papers with code

RoBLEURT Submission for WMT2021 Metrics Task

no code implementations WMT (EMNLP) 2021 Yu Wan, Dayiheng Liu, Baosong Yang, Tianchi Bi, Haibo Zhang, Boxing Chen, Weihua Luo, Derek F. Wong, Lidia S. Chao

After investigating the recent advances of trainable metrics, we conclude several aspects of vital importance to obtain a well-performed metric model by: 1) jointly leveraging the advantages of source-included model and reference-only model, 2) continuously pre-training the model with massive synthetic data pairs, and 3) fine-tuning the model with data denoising strategy.

Denoising

Challenges of Neural Machine Translation for Short Texts

no code implementations CL (ACL) 2022 Yu Wan, Baosong Yang, Derek Fai Wong, Lidia Sam Chao, Liang Yao, Haibo Zhang, Boxing Chen

After empirically investigating the rationale behind this, we summarize two challenges in NMT for STs associated with translation error types above, respectively: (1) the imbalanced length distribution in training set intensifies model inference calibration over STs, leading to more over-translation cases on STs; and (2) the lack of contextual information forces NMT to have higher data uncertainty on short sentences, and thus NMT model is troubled by considerable mistranslation errors.

Machine Translation NMT +2

GCPG: A General Framework for Controllable Paraphrase Generation

no code implementations Findings (ACL) 2022 Kexin Yang, Dayiheng Liu, Wenqiang Lei, Baosong Yang, Haibo Zhang, Xue Zhao, Wenqing Yao, Boxing Chen

Under GCPG, we reconstruct commonly adopted lexical condition (i. e., Keywords) and syntactical conditions (i. e., Part-Of-Speech sequence, Constituent Tree, Masked Template and Sentential Exemplar) and study the combination of the two types.

Paraphrase Generation Sentence

Versatile Defense Against Adversarial Attacks on Image Recognition

no code implementations13 Mar 2024 Haibo Zhang, Zhihua Yao, Kouichi Sakurai

When facing the PGD attack and the MI-FGSM attack, versatile defense model even outperforms the attack-specific models trained based on these two attacks.

Image-to-Image Translation

Efficient All-reduce for Distributed DNN Training in Optical Interconnect System

no code implementations22 Jul 2022 Fei Dai, Yawen Chen, Zhiyi Huang, Haibo Zhang, Fangfang Zhang

Our results also show that WRHT can reduce the communication time of all-reduce operation by 92. 42% and 91. 31% compared to two existing all-reduce algorithms running in the electrical interconnect system.

RoBLEURT Submission for the WMT2021 Metrics Task

no code implementations28 Apr 2022 Yu Wan, Dayiheng Liu, Baosong Yang, Tianchi Bi, Haibo Zhang, Boxing Chen, Weihua Luo, Derek F. Wong, Lidia S. Chao

After investigating the recent advances of trainable metrics, we conclude several aspects of vital importance to obtain a well-performed metric model by: 1) jointly leveraging the advantages of source-included model and reference-only model, 2) continuously pre-training the model with massive synthetic data pairs, and 3) fine-tuning the model with data denoising strategy.

Denoising

Frequency-Aware Contrastive Learning for Neural Machine Translation

no code implementations29 Dec 2021 Tong Zhang, Wei Ye, Baosong Yang, Long Zhang, Xingzhang Ren, Dayiheng Liu, Jinan Sun, Shikun Zhang, Haibo Zhang, Wen Zhao

Inspired by the observation that low-frequency words form a more compact embedding space, we tackle this challenge from a representation learning perspective.

Contrastive Learning Machine Translation +3

KGR^4: Retrieval, Retrospect, Refine and Rethink for Commonsense Generation

1 code implementation15 Dec 2021 Xin Liu, Dayiheng Liu, Baosong Yang, Haibo Zhang, Junwei Ding, Wenqing Yao, Weihua Luo, Haiying Zhang, Jinsong Su

Generative commonsense reasoning requires machines to generate sentences describing an everyday scenario given several concepts, which has attracted much attention recently.

Retrieval Sentence

Leveraging Advantages of Interactive and Non-Interactive Models for Vector-Based Cross-Lingual Information Retrieval

no code implementations3 Nov 2021 Linlong Xu, Baosong Yang, Xiaoyu Lv, Tianchi Bi, Dayiheng Liu, Haibo Zhang

Interactive and non-interactive model are the two de-facto standard frameworks in vector-based cross-lingual information retrieval (V-CLIR), which embed queries and documents in synchronous and asynchronous fashions, respectively.

Computational Efficiency Cross-Lingual Information Retrieval +4

Accelerating Fully Connected Neural Network on Optical Network-on-Chip (ONoC)

no code implementations30 Sep 2021 Fei Dai, Yawen Chen, Haibo Zhang, Zhiyi Huang

Compared with ENoC, simulation results show that under batch sizes of 64 and 128, on average ONoC can achieve 21. 02% and 12. 95% on reducing training time with 47. 85% and 39. 27% on saving energy, respectively.

Multi-Hop Transformer for Document-Level Machine Translation

no code implementations NAACL 2021 Long Zhang, Tong Zhang, Haibo Zhang, Baosong Yang, Wei Ye, Shikun Zhang

Document-level neural machine translation (NMT) has proven to be of profound value for its effectiveness on capturing contextual information.

Document Level Machine Translation Document Translation +4

Exploiting Neural Query Translation into Cross Lingual Information Retrieval

no code implementations26 Oct 2020 Liang Yao, Baosong Yang, Haibo Zhang, Weihua Luo, Boxing Chen

As a crucial role in cross-language information retrieval (CLIR), query translation has three main challenges: 1) the adequacy of translation; 2) the lack of in-domain parallel training data; and 3) the requisite of low latency.

Cross-Lingual Information Retrieval Data Augmentation +5

Self-Paced Learning for Neural Machine Translation

1 code implementation EMNLP 2020 Yu Wan, Baosong Yang, Derek F. Wong, Yikai Zhou, Lidia S. Chao, Haibo Zhang, Boxing Chen

Recent studies have proven that the training of neural machine translation (NMT) can be facilitated by mimicking the learning process of humans.

Machine Translation NMT +2

Cannot find the paper you are looking for? You can Submit a new open access paper.