Search Results for author: Si-Qing Chen

Found 17 papers, 9 papers with code

On Meta-Prompting

2 code implementations11 Dec 2023 Adrian de Wynter, Xun Wang, Qilong Gu, Si-Qing Chen

We call these approaches meta-prompting, or prompting to obtain prompts.

SCALE: Synergized Collaboration of Asymmetric Language Translation Engines

1 code implementation29 Sep 2023 Xin Cheng, Xun Wang, Tao Ge, Si-Qing Chen, Furu Wei, Dongyan Zhao, Rui Yan

In this paper, we introduce SCALE, a collaborative framework that connects compact Specialized Translation Models (STMs) and general-purpose Large Language Models (LLMs) as one unified translation engine.

Continual Learning Translation

A User-Centered Evaluation of Spanish Text Simplification

1 code implementation15 Aug 2023 Adrian de Wynter, Anthony Hevia, Si-Qing Chen

We present an evaluation of text simplification (TS) in Spanish for a production system, by means of two corpora focused in both complex-sentence and complex-word identification.

Complex Word Identification Sentence +1

In-context Autoencoder for Context Compression in a Large Language Model

1 code implementation13 Jul 2023 Tao Ge, Jing Hu, Lei Wang, Xun Wang, Si-Qing Chen, Furu Wei

We propose the In-context Autoencoder (ICAE), leveraging the power of a large language models (LLM) to compress a long context into short compact memory slots that can be directly conditioned on by the LLM for various purposes.

Language Modelling Large Language Model +3

Interactive Editing for Text Summarization

1 code implementation5 Jun 2023 Yujia Xie, Xun Wang, Si-Qing Chen, Wayne Xiong, Pengcheng He

Summarizing lengthy documents is a common and essential task in our daily lives.

Text Summarization

Joint Repetition Suppression and Content Moderation of Large Language Models

no code implementations20 Apr 2023 Minghui Zhang, Alex Sokolov, Weixin Cai, Si-Qing Chen

Natural language generation (NLG) is one of the most impactful fields in NLP, and recent years have witnessed its evolution brought about by large language models (LLMs).

Text Generation

An Evaluation on Large Language Model Outputs: Discourse and Memorization

no code implementations17 Apr 2023 Adrian de Wynter, Xun Wang, Alex Sokolov, Qilong Gu, Si-Qing Chen

We present an empirical evaluation of various outputs generated by nine of the most widely-available large language models (LLMs).

counterfactual Language Modelling +2

Semiparametric Language Models Are Scalable Continual Learners

no code implementations2 Mar 2023 Guangyue Peng, Tao Ge, Si-Qing Chen, Furu Wei, Houfeng Wang

We demonstrate that SeMem improves the scalability of semiparametric LMs for continual learning over streaming data in two ways: (1) data-wise scalability: as the model becomes stronger through continual learning, it will encounter fewer difficult cases that need to be memorized, causing the growth of the non-parametric memory to slow down over time rather than growing at a linear rate with the size of training data; (2) model-wise scalability: SeMem allows a larger model to memorize fewer samples than its smaller counterpart because it is rarer for a larger model to encounter incomprehensible cases, resulting in a non-parametric memory that does not scale linearly with model size.

Continual Learning Language Modelling +1

Pay Attention to Your Tone: Introducing a New Dataset for Polite Language Rewrite

no code implementations20 Dec 2022 Xun Wang, Tao Ge, Allen Mao, Yuki Li, Furu Wei, Si-Qing Chen

We introduce \textsc{PoliteRewrite} -- a dataset for polite language rewrite which is a novel sentence rewrite task.

Sentence Style Transfer +1

Momentum Calibration for Text Generation

no code implementations8 Dec 2022 Xingxing Zhang, Yiran Liu, Xun Wang, Pengcheng He, Yang Yu, Si-Qing Chen, Wayne Xiong, Furu Wei

The input and output of most text generation tasks can be transformed to two sequences of tokens and they can be modeled using sequence-to-sequence learning modeling tools such as Transformers.

Abstractive Text Summarization Text Generation

Latent Prompt Tuning for Text Summarization

no code implementations3 Nov 2022 Yubo Zhang, Xingxing Zhang, Xun Wang, Si-Qing Chen, Furu Wei

In this paper, we propose Lotus (shorthand for Latent Prompt Tuning for Summarization), which is a single model that can be applied in both controlled and uncontrolled (without control signals) modes.

Contrastive Learning Text Summarization

Lossless Acceleration for Seq2seq Generation with Aggressive Decoding

2 code implementations20 May 2022 Tao Ge, Heming Xia, Xin Sun, Si-Qing Chen, Furu Wei

We study lossless acceleration for seq2seq generation with a novel decoding algorithm -- Aggressive Decoding.

Abstractive Text Summarization Grammatical Error Correction +4

Speculative Decoding: Exploiting Speculative Execution for Accelerating Seq2seq Generation

2 code implementations30 Mar 2022 Heming Xia, Tao Ge, Peiyi Wang, Si-Qing Chen, Furu Wei, Zhifang Sui

We propose Speculative Decoding (SpecDec), for the first time ever, to formally study exploiting the idea of speculative execution to accelerate autoregressive (AR) decoding.

Abstractive Text Summarization Machine Translation +1

EdgeFormer: A Parameter-Efficient Transformer for On-Device Seq2seq Generation

1 code implementation16 Feb 2022 Tao Ge, Si-Qing Chen, Furu Wei

We introduce EdgeFormer -- a parameter-efficient Transformer for on-device seq2seq generation under the strict computation and memory constraints.

Grammatical Error Correction Knowledge Distillation +2

Cannot find the paper you are looking for? You can Submit a new open access paper.