Search Results for author: Shaolei Zhang

Found 22 papers, 15 papers with code

Truth-Aware Context Selection: Mitigating the Hallucinations of Large Language Models Being Misled by Untruthful Contexts

1 code implementation12 Mar 2024 Tian Yu, Shaolei Zhang, Yang Feng

Although large language models (LLMs) have demonstrated impressive text generation capabilities, they are easily misled by the untruthful context provided by users or knowledge augmentation tools, thereby producing hallucinations.

Text Generation

TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space

1 code implementation27 Feb 2024 Shaolei Zhang, Tian Yu, Yang Feng

During inference, by editing LLM's internal representations in truthful space, TruthX effectively enhances the truthfulness of LLMs.

Contrastive Learning Hallucination +5

SiLLM: Large Language Models for Simultaneous Machine Translation

1 code implementation20 Feb 2024 Shoutao Guo, Shaolei Zhang, Zhengrui Ma, Min Zhang, Yang Feng

We propose SiLLM, which delegates the two sub-tasks to separate agents, thereby incorporating LLM into SiMT.

Machine Translation Sentence +1

Unified Segment-to-Segment Framework for Simultaneous Sequence Generation

no code implementations NeurIPS 2023 Shaolei Zhang, Yang Feng

To accomplish this, Seg2Seg introduces a latent segment as the pivot between source to target and explores all potential source-target mappings via the proposed expectation training, thereby learning the optimal moments for generating.

Machine Translation Multi-Task Learning +3

Non-autoregressive Streaming Transformer for Simultaneous Translation

1 code implementation23 Oct 2023 Zhengrui Ma, Shaolei Zhang, Shoutao Guo, Chenze Shao, Min Zhang, Yang Feng

Simultaneous machine translation (SiMT) models are trained to strike a balance between latency and translation quality.

Machine Translation Translation

Simultaneous Machine Translation with Tailored Reference

no code implementations20 Oct 2023 Shoutao Guo, Shaolei Zhang, Yang Feng

Training the model with ground-truth at low latency may introduce forced anticipations, whereas utilizing reference consistent with the source word order at high latency results in performance degradation.

Machine Translation Sentence +1

Glancing Future for Simultaneous Machine Translation

1 code implementation12 Sep 2023 Shoutao Guo, Shaolei Zhang, Yang Feng

Simultaneous machine translation (SiMT) outputs translation while reading the source sentence.

Machine Translation Sentence +1

BayLing: Bridging Cross-lingual Alignment and Instruction Following through Interactive Translation for Large Language Models

1 code implementation19 Jun 2023 Shaolei Zhang, Qingkai Fang, Zhuocheng Zhang, Zhengrui Ma, Yan Zhou, Langlin Huang, Mengyu Bu, Shangtong Gui, Yunji Chen, Xilin Chen, Yang Feng

To minimize human workload, we propose to transfer the capabilities of language generation and instruction following from English to other languages through an interactive translation task.

Instruction Following Text Generation +1

End-to-End Simultaneous Speech Translation with Differentiable Segmentation

1 code implementation25 May 2023 Shaolei Zhang, Yang Feng

Therefore, learning to segment the speech inputs at those moments that are beneficial for the translation model to produce high-quality translation is the key to SimulST.

Segmentation Translation

Learning Optimal Policy for Simultaneous Machine Translation via Binary Search

1 code implementation22 May 2023 Shoutao Guo, Shaolei Zhang, Yang Feng

Simultaneous machine translation (SiMT) starts to output translation while reading the source sentence and needs a precise policy to decide when to output the generated translation.

Machine Translation Sentence +1

Hidden Markov Transformer for Simultaneous Machine Translation

1 code implementation1 Mar 2023 Shaolei Zhang, Yang Feng

Simultaneous machine translation (SiMT) outputs the target sequence while receiving the source sequence, and hence learning when to start translating each target token is the core challenge for SiMT task.

Machine Translation Translation

TriDoNet: A Triple Domain Model-driven Network for CT Metal Artifact Reduction

no code implementations14 Nov 2022 Baoshun Shi, Ke Jiang, Shaolei Zhang, Qiusheng Lian, Yanwei Qin

Recent deep learning-based methods have achieved promising performance for computed tomography metal artifact reduction (CTMAR).

Contrastive Learning Metal Artifact Reduction

Information-Transport-based Policy for Simultaneous Translation

1 code implementation22 Oct 2022 Shaolei Zhang, Yang Feng

Simultaneous translation (ST) outputs translation while receiving the source inputs, and hence requires a policy to determine whether to translate a target token or wait for the next source token.

Machine Translation Translation

Turning Fixed to Adaptive: Integrating Post-Evaluation into Simultaneous Machine Translation

1 code implementation21 Oct 2022 Shoutao Guo, Shaolei Zhang, Yang Feng

Compared to the fixed policy, the adaptive policy achieves better latency-quality tradeoffs by adopting a flexible translation policy.

Machine Translation Sentence +1

Reducing Position Bias in Simultaneous Machine Translation with Length-Aware Framework

no code implementations ACL 2022 Shaolei Zhang, Yang Feng

Simultaneous machine translation (SiMT) starts translating while receiving the streaming source inputs, and hence the source sentence is always incomplete during translating.

Machine Translation Position +2

Gaussian Multi-head Attention for Simultaneous Machine Translation

1 code implementation Findings (ACL) 2022 Shaolei Zhang, Yang Feng

For SiMT policy, GMA models the aligned source position of each target word, and accordingly waits until its aligned position to start translating.

Machine Translation Position +1

Modeling Dual Read/Write Paths for Simultaneous Machine Translation

1 code implementation ACL 2022 Shaolei Zhang, Yang Feng

According to duality constraints, the read/write path in source-to-target and target-to-source SiMT models can be mapped to each other.

Machine Translation Sentence +1

Modeling Concentrated Cross-Attention for Neural Machine Translation with Gaussian Mixture Model

no code implementations Findings (EMNLP) 2021 Shaolei Zhang, Yang Feng

Cross-attention is an important component of neural machine translation (NMT), which is always realized by dot-product attention in previous methods.

Machine Translation NMT +2

Universal Simultaneous Machine Translation with Mixture-of-Experts Wait-k Policy

1 code implementation EMNLP 2021 Shaolei Zhang, Yang Feng

Simultaneous machine translation (SiMT) generates translation before reading the entire source sentence and hence it has to trade off between translation quality and latency.

Machine Translation Sentence +1

Future-Guided Incremental Transformer for Simultaneous Translation

no code implementations23 Dec 2020 Shaolei Zhang, Yang Feng, Liangyou Li

Simultaneous translation (ST) starts translations synchronously while reading source sentences, and is used in many online scenarios.

Knowledge Distillation Translation

Cannot find the paper you are looking for? You can Submit a new open access paper.