Paraphrase Generation

68 papers with code • 3 benchmarks • 16 datasets

Paraphrase Generation involves transforming a natural language sentence to a new sentence, that has the same semantic meaning but a different syntactic or lexical surface form.

TESS: Text-to-Text Self-Conditioned Simplex Diffusion

allenai/tess-diffusion 15 May 2023

Diffusion models have emerged as a powerful paradigm for generation, obtaining strong performance in various continuous domains.

11
15 May 2023

Lost in Translationese? Reducing Translation Effect Using Abstract Meaning Representation

shirawein/amr-translationese 23 Apr 2023

Though individual translated texts are often fluent and preserve meaning, at a large scale, translated texts have statistical tendencies which distinguish them from text originally written in the language ("translationese") and can affect model performance.

0
23 Apr 2023

Paraphrasing evades detectors of AI-generated text, but retrieval is an effective defense

martiansideofthemoon/ai-detection-paraphrases NeurIPS 2023

To increase the robustness of AI-generated text detection to paraphrase attacks, we introduce a simple defense that relies on retrieving semantically-similar generations and must be maintained by a language model API provider.

110
23 Mar 2023

kNN-BOX: A Unified Framework for Nearest Neighbor Generation

njunlp/knn-box 27 Feb 2023

Augmenting the base neural model with a token-level symbolic datastore is a novel generation paradigm and has achieved promising results in machine translation (MT).

99
27 Feb 2023

Syntactically Robust Training on Partially-Observed Data for Open Information Extraction

qijimrc/robustoie 17 Jan 2023

In this paper, we propose a syntactically robust training framework that enables models to be trained on a syntactic-abundant distribution based on diverse paraphrase generation.

6
17 Jan 2023

Language as a Latent Sequence: deep latent variable models for semi-supervised paraphrase generation

jialin-yu/latent-sequence-paraphrase 5 Jan 2023

To leverage information from text pairs, we additionally introduce a novel supervised model we call dual directional learning (DDL), which is designed to integrate with our proposed VSAR model.

0
05 Jan 2023

How Large Language Models are Transforming Machine-Paraphrased Plagiarism

jpwahle/emnlp22-transforming 7 Oct 2022

The recent success of large language models for text generation poses a severe threat to academic integrity, as plagiarists can generate realistic paraphrases indistinguishable from original work.

10
07 Oct 2022

Continuous Decomposition of Granularity for Neural Paraphrase Generation

guxd/c-dnpg COLING 2022

While Transformers have had significant success in paragraph generation, they treat sentences as linear sequences of tokens and often neglect their hierarchical information.

8
05 Sep 2022

PCC: Paraphrasing with Bottom-k Sampling and Cyclic Learning for Curriculum Data Augmentation

hongyuanluke/pcc 17 Aug 2022

This paper presents \textbf{PCC}: \textbf{P}araphrasing with Bottom-k Sampling and \textbf{C}yclic Learning for \textbf{C}urriculum Data Augmentation, a novel CDA framework via paraphrasing, which exploits the textual paraphrase similarity as the curriculum difficulty measure.

1
17 Aug 2022

'John ate 5 apples' != 'John ate some apples': Self-Supervised Paraphrase Quality Detection for Algebraic Word Problems

ads-ai/paraqd 16 Jun 2022

There is a need for paraphrase scoring methods in the context of AWP to enable the training of good paraphrasers.

2
16 Jun 2022