Paraphrase Generation

66 papers with code • 3 benchmarks • 16 datasets

Paraphrase Generation involves transforming a natural language sentence to a new sentence, that has the same semantic meaning but a different syntactic or lexical surface form.

Latest papers with no code

SemEval-2024 Shared Task 6: SHROOM, a Shared-task on Hallucinations and Related Observable Overgeneration Mistakes

no code yet • 12 Mar 2024

This paper presents the results of the SHROOM, a shared task focused on detecting hallucinations: outputs from natural language generation (NLG) systems that are fluent, yet inaccurate.

Fine-tuning CLIP Text Encoders with Two-step Paraphrasing

no code yet • 23 Feb 2024

Contrastive language-image pre-training (CLIP) models have demonstrated considerable success across various vision-language tasks, such as text-to-image retrieval, where the model is required to effectively process natural language input to produce an accurate visual output.

Neural Machine Translation for Malayalam Paraphrase Generation

no code yet • 31 Jan 2024

This study explores four methods of generating paraphrases in Malayalam, utilizing resources available for English paraphrasing and pre-trained Neural Machine Translation (NMT) models.

Vector-Quantized Prompt Learning for Paraphrase Generation

no code yet • 25 Nov 2023

Therefore, we present vector-quantized prompts as the cues to control the generation of pre-trained models.

Contextual Data Augmentation for Task-Oriented Dialog Systems

no code yet • 16 Oct 2023

While dialog response generation has been widely studied on the agent side, it is not evident if similar generative models can be used to generate a large variety of, and often unexpected, user inputs that real dialog systems encounter in practice.

Automatic and Human-AI Interactive Text Generation

no code yet • 5 Oct 2023

In this tutorial, we focus on text-to-text generation, a class of natural language generation (NLG) tasks, that takes a piece of text as input and then generates a revision that is improved according to some specific criteria (e. g., readability or linguistic styles), while largely retaining the original meaning and the length of the text.

Emotion and Sentiment Guided Paraphrasing

no code yet • 8 Jun 2023

Paraphrase generation, a. k. a.

Impossible Distillation: from Low-Quality Model to High-Quality Dataset & Model for Summarization and Paraphrasing

no code yet • 26 May 2023

We present Impossible Distillation, a novel framework for paraphrasing and sentence summarization, that distills a high-quality dataset and model from a low-quality teacher that itself cannot perform these tasks.

Coherence and Diversity through Noise: Self-Supervised Paraphrase Generation via Structure-Aware Denoising

no code yet • 6 Feb 2023

This allows for learning a denoising function that operates over both aspects and produces semantically equivalent and syntactically diverse outputs through grounded noise injection.

Unsupervised Syntactically Controlled Paraphrase Generation with Abstract Meaning Representations

no code yet • 2 Nov 2022

In this paper, we demonstrate that leveraging Abstract Meaning Representations (AMR) can greatly improve the performance of unsupervised syntactically controlled paraphrase generation.