Search Results for author: Jiali Zeng

Found 19 papers, 11 papers with code

Tencent Translation System for the WMT21 News Translation Task

no code implementations WMT (EMNLP) 2021 Longyue Wang, Mu Li, Fangxu Liu, Shuming Shi, Zhaopeng Tu, Xing Wang, Shuangzhi Wu, Jiali Zeng, Wen Zhang

Based on our success in the last WMT, we continuously employed advanced techniques such as large batch training, data selection and data filtering.

Data Augmentation Translation

Recurrent Attention for Neural Machine Translation

1 code implementation EMNLP 2021 Jiali Zeng, Shuangzhi Wu, Yongjing Yin, Yufan Jiang, Mu Li

Across an extensive set of experiments on 10 machine translation tasks, we find that RAN models are competitive and outperform their Transformer counterpart in certain scenarios, with fewer parameters and inference time.

Machine Translation NMT +1

Generative Multi-Modal Knowledge Retrieval with Large Language Models

no code implementations16 Jan 2024 Xinwei Long, Jiali Zeng, Fandong Meng, Zhiyuan Ma, Kaiyan Zhang, BoWen Zhou, Jie zhou

Knowledge retrieval with multi-modal queries plays a crucial role in supporting knowledge-intensive multi-modal applications.

Retrieval

Improving Machine Translation with Large Language Models: A Preliminary Study with Cooperative Decoding

no code implementations6 Nov 2023 Jiali Zeng, Fandong Meng, Yongjing Yin, Jie zhou

Contemporary translation engines built upon the encoder-decoder framework have reached a high level of development, while the emergence of Large Language Models (LLMs) has disrupted their position by offering the potential for achieving superior translation quality.

Machine Translation NMT +1

TIM: Teaching Large Language Models to Translate with Comparison

1 code implementation10 Jul 2023 Jiali Zeng, Fandong Meng, Yongjing Yin, Jie zhou

Open-sourced large language models (LLMs) have demonstrated remarkable efficacy in various tasks with instruction tuning.

Translation

Soft Language Clustering for Multilingual Model Pre-training

no code implementations13 Jun 2023 Jiali Zeng, Yufan Jiang, Yongjing Yin, Yi Jing, Fandong Meng, Binghuai Lin, Yunbo Cao, Jie zhou

Multilingual pre-trained language models have demonstrated impressive (zero-shot) cross-lingual transfer abilities, however, their performance is hindered when the target language has distant typology from source languages or when pre-training data is limited in size.

Clustering Question Answering +5

DualNER: A Dual-Teaching framework for Zero-shot Cross-lingual Named Entity Recognition

no code implementations15 Nov 2022 Jiali Zeng, Yufan Jiang, Yongjing Yin, Xu Wang, Binghuai Lin, Yunbo Cao

We present DualNER, a simple and effective framework to make full use of both annotated source language corpus and unlabeled target language text for zero-shot cross-lingual named entity recognition (NER).

named-entity-recognition Named Entity Recognition +1

Contrastive Learning with Prompt-derived Virtual Semantic Prototypes for Unsupervised Sentence Embedding

no code implementations7 Nov 2022 Jiali Zeng, Yongjing Yin, Yufan Jiang, Shuangzhi Wu, Yunbo Cao

Specifically, with the help of prompts, we construct virtual semantic prototypes to each instance, and derive negative prototypes by using the negative form of the prompts.

Clustering Contrastive Learning +5

An Efficient Coarse-to-Fine Facet-Aware Unsupervised Summarization Framework based on Semantic Blocks

1 code implementation COLING 2022 Xinnian Liang, Jing Li, Shuangzhi Wu, Jiali Zeng, Yufan Jiang, Mu Li, Zhoujun Li

To tackle this problem, in this paper, we proposed an efficient Coarse-to-Fine Facet-Aware Ranking (C2F-FAR) framework for unsupervised long document summarization, which is based on the semantic block.

Document Summarization

Task-guided Disentangled Tuning for Pretrained Language Models

1 code implementation Findings (ACL) 2022 Jiali Zeng, Yufan Jiang, Shuangzhi Wu, Yongjing Yin, Mu Li

Pretrained language models (PLMs) trained on large-scale unlabeled corpus are typically fine-tuned on task-specific downstream datasets, which have produced state-of-the-art results on various NLP tasks.

Learning Confidence for Transformer-based Neural Machine Translation

1 code implementation ACL 2022 Yu Lu, Jiali Zeng, Jiajun Zhang, Shuangzhi Wu, Mu Li

Confidence estimation aims to quantify the confidence of the model prediction, providing an expectation of success.

Machine Translation NMT +2

Type-Driven Multi-Turn Corrections for Grammatical Error Correction

1 code implementation Findings (ACL) 2022 Shaopeng Lai, Qingyu Zhou, Jiali Zeng, Zhongli Li, Chao Li, Yunbo Cao, Jinsong Su

First, they simply mix additionally-constructed training instances and original ones to train models, which fails to help models be explicitly aware of the procedure of gradual corrections.

Data Augmentation Grammatical Error Correction +1

Attention Calibration for Transformer in Neural Machine Translation

no code implementations ACL 2021 Yu Lu, Jiali Zeng, Jiajun Zhang, Shuangzhi Wu, Mu Li

Attention mechanisms have achieved substantial improvements in neural machine translation by dynamically selecting relevant inputs for different predictions.

Machine Translation Translation

Neural Simile Recognition with Cyclic Multitask Learning and Local Attention

1 code implementation19 Dec 2019 Jiali Zeng, Linfeng Song, Jinsong Su, Jun Xie, Wei Song, Jiebo Luo

Simile recognition is to detect simile sentences and to extract simile components, i. e., tenors and vehicles.

Sentence Sentence Classification

Graph-based Neural Sentence Ordering

1 code implementation16 Dec 2019 Yongjing Yin, Linfeng Song, Jinsong Su, Jiali Zeng, Chulun Zhou, Jiebo Luo

Sentence ordering is to restore the original paragraph from a set of sentences.

Sentence Sentence Ordering

Iterative Dual Domain Adaptation for Neural Machine Translation

no code implementations IJCNLP 2019 Jiali Zeng, Yang Liu, Jinsong Su, Yubin Ge, Yaojie Lu, Yongjing Yin, Jiebo Luo

Previous studies on the domain adaptation for neural machine translation (NMT) mainly focus on the one-pass transferring out-of-domain translation knowledge to in-domain NMT model.

Domain Adaptation Knowledge Distillation +4

Cannot find the paper you are looking for? You can Submit a new open access paper.