Search Results for author: Dongyan Zhao

Found 158 papers, 70 papers with code

Finding the Dominant Winning Ticket in Pre-Trained Language Models

no code implementations Findings (ACL) 2022 Zhuocheng Gong, Di He, Yelong Shen, Tie-Yan Liu, Weizhu Chen, Dongyan Zhao, Ji-Rong Wen, Rui Yan

Empirically, we show that (a) the dominant winning ticket can achieve performance that is comparable with that of the full-parameter model, (b) the dominant winning ticket is transferable across different tasks, (c) and the dominant winning ticket has a natural structure within each parameter matrix.

Combining Curriculum Learning and Knowledge Distillation for Dialogue Generation

no code implementations Findings (EMNLP) 2021 Qingqing Zhu, Xiuying Chen, Pengfei Wu, Junfei Liu, Dongyan Zhao

Hence, in this paper, we introduce a combination of curriculum learning and knowledge distillation for efficient dialogue generation models, where curriculum learning can help knowledge distillation from data and model aspects.

Dialogue Generation Knowledge Distillation +1

Understanding Procedural Text using Interactive Entity Networks

no code implementations EMNLP 2020 Jizhi Tang, Yansong Feng, Dongyan Zhao

Recent efforts have made great progress to track multiple entities in a procedural text, but usually treat each entity separately and ignore the fact that there are often multiple entities interacting with each other during one process, some of which are even explicitly mentioned.

Reading Comprehension

Reciprocal Learning of Knowledge Retriever and Response Ranker for Knowledge-Grounded Conversations

no code implementations COLING 2022 Jiazhan Feng, Chongyang Tao, Zhen Li, Chang Liu, Tao Shen, Dongyan Zhao

In this paper, we propose a reciprocal learning approach to jointly optimize a knowledge retriever and a response ranker for knowledge-grounded response retrieval without ground-truth knowledge labels.

Retrieval

Combining Impression Feature Representation for Multi-turn Conversational Question Answering

no code implementations CCL 2020 Shaoling Jing, Shibo Hong, Dongyan Zhao, Haihua Xie, Zhi Tang

Multi-turn conversational Question Answering (ConvQA) is a practical task that requires the understanding of conversation history, such as previous QA pairs, the passage context, and current question.

Conversational Question Answering feature selection

Plan-CVAE: A Planning-based Conditional Variational Autoencoder for Story Generation

no code implementations CCL 2020 Lin Wang, Juntao Li, Rui Yan, Dongyan Zhao

Story generation is a challenging task of automatically creating natural languages to describe a sequence of events, which requires outputting text with not only a consistent topic but also novel wordings.

Story Generation

ProphetChat: Enhancing Dialogue Generation with Simulation of Future Conversation

no code implementations ACL 2022 Chang Liu, Xu Tan, Chongyang Tao, Zhenxin Fu, Dongyan Zhao, Tie-Yan Liu, Rui Yan

To enable the chatbot to foresee the dialogue future, we design a beam-search-like roll-out strategy for dialogue future simulation using a typical dialogue generation model and a dialogue selector.

Dialogue Generation Response Generation

StyleChat: Learning Recitation-Augmented Memory in LLMs for Stylized Dialogue Generation

no code implementations18 Mar 2024 Jinpeng Li, Zekai Zhang, Quan Tu, Xin Cheng, Dongyan Zhao, Rui Yan

Furthermore, although many prompt-based methods have been proposed to accomplish specific tasks, their performance in complex real-world scenarios involving a wide variety of dialog styles further enhancement.

Dialogue Generation

HawkEye: Training Video-Text LLMs for Grounding Text in Videos

1 code implementation15 Mar 2024 Yueqian Wang, Xiaojun Meng, Jianxin Liang, Yuxuan Wang, Qun Liu, Dongyan Zhao

Video-text Large Language Models (video-text LLMs) have shown remarkable performance in answering questions and holding conversations on simple videos.

Video Grounding Video Question Answering

What Makes Quantization for Large Language Models Hard? An Empirical Study from the Lens of Perturbation

no code implementations11 Mar 2024 Zhuocheng Gong, Jiahao Liu, Jingang Wang, Xunliang Cai, Dongyan Zhao, Rui Yan

Our findings reveal several connections between the properties of perturbations and LLM performance, providing insights into the failure cases of uniform quantization and suggesting potential solutions to improve the robustness of LLM quantization.

Computational Efficiency Quantization

PPTC-R benchmark: Towards Evaluating the Robustness of Large Language Models for PowerPoint Task Completion

1 code implementation6 Mar 2024 Zekai Zhang, Yiduo Guo, Yaobo Liang, Dongyan Zhao, Nan Duan

The growing dependence on Large Language Models (LLMs) for finishing user instructions necessitates a comprehensive understanding of their robustness to complex task completion in real-world situations.

Sentence

Probing Multimodal Large Language Models for Global and Local Semantic Representations

1 code implementation27 Feb 2024 Mingxu Tao, Quzhe Huang, Kun Xu, Liwei Chen, Yansong Feng, Dongyan Zhao

The advancement of Multimodal Large Language Models (MLLMs) has greatly accelerated the development of applications in understanding integrated texts and images.

object-detection Object Detection +1

Chain-of-Discussion: A Multi-Model Framework for Complex Evidence-Based Question Answering

no code implementations26 Feb 2024 Mingxu Tao, Dongyan Zhao, Yansong Feng

Open-ended question answering requires models to find appropriate evidence to form well-reasoned, comprehensive and helpful answers.

Evidence Selection Open-Ended Question Answering +1

LSTP: Language-guided Spatial-Temporal Prompt Learning for Long-form Video-Text Understanding

1 code implementation25 Feb 2024 Yuxuan Wang, Yueqian Wang, Pengfei Wu, Jianxin Liang, Dongyan Zhao, Zilong Zheng

Despite progress in video-language modeling, the computational challenge of interpreting long-form videos in response to task-specific linguistic queries persists, largely due to the complexity of high-dimensional video data and the misalignment between language and visual cues over space and time.

Computational Efficiency Language Modelling +3

STAIR: Spatial-Temporal Reasoning with Auditable Intermediate Results for Video Question Answering

1 code implementation8 Jan 2024 Yueqian Wang, Yuxuan Wang, Kai Chen, Dongyan Zhao

However, most models can only handle simple videos in terms of temporal reasoning, and their performance tends to drop when answering temporal-reasoning questions on long and informative videos.

Question Answering Video Question Answering

Multi-Granularity Information Interaction Framework for Incomplete Utterance Rewriting

no code implementations19 Dec 2023 Haowei Du, Dinghao Zhang, Chen Li, Yang Li, Dongyan Zhao

Recent approaches in Incomplete Utterance Rewriting (IUR) fail to capture the source of important words, which is crucial to edit the incomplete utterance, and introduce words from irrelevant utterances.

Relation-Aware Question Answering for Heterogeneous Knowledge Graphs

1 code implementation19 Dec 2023 Haowei Du, Quzhe Huang, Chen Li, Chen Zhang, Yang Li, Dongyan Zhao

To address this issue, we construct a \textbf{dual relation graph} where each node denotes a relation in the original KG (\textbf{primal entity graph}) and edges are constructed between relations sharing same head or tail entities.

Knowledge Base Question Answering Knowledge Graphs +1

A Step Closer to Comprehensive Answers: Constrained Multi-Stage Question Decomposition with Large Language Models

1 code implementation13 Nov 2023 Hejing Cao, Zhenwei An, Jiazhan Feng, Kun Xu, Liwei Chen, Dongyan Zhao

While large language models exhibit remarkable performance in the Question Answering task, they are susceptible to hallucinations.

Question Answering

Language Models can be Logical Solvers

no code implementations10 Nov 2023 Jiazhan Feng, Ruochen Xu, Junheng Hao, Hiteshi Sharma, Yelong Shen, Dongyan Zhao, Weizhu Chen

Despite their impressive performance, any parsing errors will inevitably result in the failure of the execution of the external logical solver and no answer to the logical questions.

Decision Making Language Modelling +1

PPTC Benchmark: Evaluating Large Language Models for PowerPoint Task Completion

1 code implementation3 Nov 2023 Yiduo Guo, Zekai Zhang, Yaobo Liang, Dongyan Zhao, Nan Duan

Recent evaluations of Large Language Models (LLMs) have centered around testing their zero-shot/few-shot capabilities for basic natural language tasks and their ability to translate instructions into tool APIs.

Improving Input-label Mapping with Demonstration Replay for In-context Learning

no code implementations30 Oct 2023 Zhuocheng Gong, Jiahao Liu, Qifan Wang, Jingang Wang, Xunliang Cai, Dongyan Zhao, Rui Yan

The effectiveness of ICL can be attributed to the strong language modeling capabilities of large language models (LLMs), which enable them to learn the mapping between input and labels based on in-context demonstrations.

In-Context Learning Language Modelling

From Simple to Complex: A Progressive Framework for Document-level Informative Argument Extraction

1 code implementation25 Oct 2023 Quzhe Huang, Yanxi Zhang, Dongyan Zhao

These methods extract events according to their appearance order in the document, however, the event that appears in the first sentence does not mean that it is the easiest to extract.

Event Argument Extraction Sentence

Retrieval-based Knowledge Transfer: An Effective Approach for Extreme Large Language Model Compression

no code implementations24 Oct 2023 Jiduan Liu, Jiahao Liu, Qifan Wang, Jingang Wang, Xunliang Cai, Dongyan Zhao, Ran Lucien Wang, Rui Yan

In particular, our approach extracts knowledge from LLMs to construct a knowledge store, from which the small-scale model can retrieve relevant information and leverage it for effective inference.

Language Modelling Large Language Model +3

SCALE: Synergized Collaboration of Asymmetric Language Translation Engines

1 code implementation29 Sep 2023 Xin Cheng, Xun Wang, Tao Ge, Si-Qing Chen, Furu Wei, Dongyan Zhao, Rui Yan

In this paper, we introduce SCALE, a collaborative framework that connects compact Specialized Translation Models (STMs) and general-purpose Large Language Models (LLMs) as one unified translation engine.

Continual Learning Translation

Teaching Text-to-Image Models to Communicate in Dialog

no code implementations27 Sep 2023 Xiaowen Sun, Jiazhan Feng, Yuxuan Wang, Yuxuan Lai, Xingyu Shen, Dongyan Zhao

In this paper, we focus on the innovative dialog-to-image generation task, where the model synthesizes a high-resolution image aligned with the given dialog context as a response.

Sentence Text-to-Image Generation

CharacterChat: Learning towards Conversational AI with Personalized Social Support

1 code implementation20 Aug 2023 Quan Tu, Chuanqi Chen, Jinpeng Li, Yanran Li, Shuo Shang, Dongyan Zhao, Ran Wang, Rui Yan

In our modern, fast-paced, and interconnected world, the importance of mental well-being has grown into a matter of great urgency.

Leveraging Denoised Abstract Meaning Representation for Grammatical Error Correction

no code implementations5 Jul 2023 Hejing Cao, Dongyan Zhao

Experiments on the BEA-2019 shared task and the CoNLL-2014 shared task have shown that AMR-GEC performs comparably to a set of strong baselines with a large number of synthetic data.

Denoising Grammatical Error Correction

Class-Incremental Learning based on Label Generation

1 code implementation22 Jun 2023 Yijia Shao, Yiduo Guo, Dongyan Zhao, Bing Liu

Despite the great success of pre-trained language models, it is still a challenge to use these models for continual learning, especially for the class-incremental learning (CIL) setting due to catastrophic forgetting (CF).

Class Incremental Learning Incremental Learning

From the One, Judge of the Whole: Typed Entailment Graph Construction with Predicate Generation

1 code implementation7 Jun 2023 Zhibin Chen, Yansong Feng, Dongyan Zhao

Entailment Graphs (EGs) have been constructed based on extracted corpora as a strong and explainable form to indicate context-independent entailment relations in natural languages.

graph construction

MoviePuzzle: Visual Narrative Reasoning through Multimodal Order Learning

no code implementations4 Jun 2023 Jianghui Wang, Yuxuan Wang, Dongyan Zhao, Zilong Zheng

We introduce MoviePuzzle, a novel challenge that targets visual narrative reasoning and holistic movie understanding.

Benchmarking Contrastive Learning +1

How Many Answers Should I Give? An Empirical Study of Multi-Answer Reading Comprehension

1 code implementation1 Jun 2023 Chen Zhang, Jiuheng Lin, Xiao Liu, Yuxuan Lai, Yansong Feng, Dongyan Zhao

We further analyze how well different paradigms of current multi-answer MRC models deal with different types of multi-answer instances.

Machine Reading Comprehension

Shuo Wen Jie Zi: Rethinking Dictionaries and Glyphs for Chinese Language Pre-training

1 code implementation30 May 2023 Yuxuan Wang, Jianghui Wang, Dongyan Zhao, Zilong Zheng

We introduce CDBERT, a new learning paradigm that enhances the semantics understanding ability of the Chinese PLMs with dictionary knowledge and structure of Chinese characters.

Contrastive Learning

PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language Models

no code implementations30 May 2023 Zhuocheng Gong, Jiahao Liu, Qifan Wang, Yang Yang, Jingang Wang, Wei Wu, Yunsen Xian, Dongyan Zhao, Rui Yan

While transformer-based pre-trained language models (PLMs) have dominated a number of NLP applications, these models are heavy to deploy and expensive to use.

Quantization

VSTAR: A Video-grounded Dialogue Dataset for Situated Semantic Understanding with Scene and Topic Transitions

1 code implementation30 May 2023 Yuxuan Wang, Zilong Zheng, Xueliang Zhao, Jinpeng Li, Yueqian Wang, Dongyan Zhao

Video-grounded dialogue understanding is a challenging problem that requires machine to perceive, parse and reason over situated semantics extracted from weakly aligned video and dialogues.

Dialogue Generation Dialogue Understanding +2

More than Classification: A Unified Framework for Event Temporal Relation Extraction

no code implementations28 May 2023 Quzhe Huang, Yutong Hu, Shengqi Zhu, Yansong Feng, Chang Liu, Dongyan Zhao

After examining the relation definitions in various ETRE tasks, we observe that all relations can be interpreted using the start and end time points of events.

Multi-Label Classification Relation +1

RankCSE: Unsupervised Sentence Representations Learning via Learning to Rank

1 code implementation26 May 2023 Jiduan Liu, Jiahao Liu, Qifan Wang, Jingang Wang, Wei Wu, Yunsen Xian, Dongyan Zhao, Kai Chen, Rui Yan

In this paper, we propose a novel approach, RankCSE, for unsupervised sentence representation learning, which incorporates ranking consistency and ranking distillation with contrastive learning into a unified framework.

Contrastive Learning Learning-To-Rank +4

Dealing with Cross-Task Class Discrimination in Online Continual Learning

1 code implementation CVPR 2023 Yiduo Guo, Bing Liu, Dongyan Zhao

A novel optimization objective with a gradient-based adaptive method is proposed to dynamically deal with the problem in the online CL process.

Class Incremental Learning Incremental Learning

Analyzing and Reducing the Performance Gap in Cross-Lingual Transfer with Fine-tuning Slow and Fast

no code implementations19 May 2023 Yiduo Guo, Yaobo Liang, Dongyan Zhao, Bing Liu, Duan Nan

Existing research has shown that a multilingual pre-trained language model fine-tuned with one (source) language also performs well on downstream tasks for non-source languages, even though no fine-tuning is done on these languages.

Cross-Lingual Transfer Language Modelling

Decouple knowledge from parameters for plug-and-play language modeling

1 code implementation19 May 2023 Xin Cheng, Yankai Lin, Xiuying Chen, Dongyan Zhao, Rui Yan

The key intuition is to decouple the knowledge storage from model parameters with an editable and scalable key-value memory and leverage knowledge in an explainable manner by knowledge retrieval in the DPM.

Domain Adaptation Language Modelling +1

Smart Word Suggestions for Writing Assistance

1 code implementation17 May 2023 Chenshuo Wang, Shaoguang Mao, Tao Ge, Wenshan Wu, Xun Wang, Yan Xia, Jonathan Tien, Dongyan Zhao

The training dataset comprises over 3. 7 million sentences and 12. 7 million suggestions generated through rules.

Synergistic Interplay between Search and Large Language Models for Information Retrieval

1 code implementation12 May 2023 Jiazhan Feng, Chongyang Tao, Xiubo Geng, Tao Shen, Can Xu, Guodong Long, Dongyan Zhao, Daxin Jiang

Information retrieval (IR) plays a crucial role in locating relevant resources from vast amounts of data, and its applications have evolved from traditional knowledge bases to modern retrieval models (RMs).

Information Retrieval Retrieval

A Frustratingly Easy Improvement for Position Embeddings via Random Padding

no code implementations8 May 2023 Mingxu Tao, Yansong Feng, Dongyan Zhao

Since the embeddings of rear positions are updated fewer times than the front position embeddings, the rear ones may not be properly trained.

Extractive Question-Answering Position +1

Lift Yourself Up: Retrieval-augmented Text Generation with Self Memory

1 code implementation3 May 2023 Xin Cheng, Di Luo, Xiuying Chen, Lemao Liu, Dongyan Zhao, Rui Yan

In this paper, by exploring the duality of the primal problem: better generation also prompts better memory, we propose a novel framework, selfmem, which addresses this limitation by iteratively employing a retrieval-augmented generator to create an unbounded memory pool and using a memory selector to choose one output as memory for the subsequent generation round.

Abstractive Text Summarization Dialogue Generation +2

Learning to Plan with Natural Language

1 code implementation20 Apr 2023 Yiduo Guo, Yaobo Liang, Chenfei Wu, Wenshan Wu, Dongyan Zhao, Nan Duan

To obtain it, we propose the Learning to Plan method, which involves two phases: (1) In the first learning task plan phase, it iteratively updates the task plan with new step-by-step solutions and behavioral instructions, which are obtained by prompting LLMs to derive from training error feedback.

Transfer Learning

Can BERT Refrain from Forgetting on Sequential Tasks? A Probing Study

1 code implementation2 Mar 2023 Mingxu Tao, Yansong Feng, Dongyan Zhao

Large pre-trained language models help to achieve state of the art on a variety of natural language processing (NLP) tasks, nevertheless, they still suffer from forgetting when incrementally learning a sequence of tasks.

Extractive Question-Answering Incremental Learning +3

Cross-Lingual Question Answering over Knowledge Base as Reading Comprehension

1 code implementation26 Feb 2023 Chen Zhang, Yuxuan Lai, Yansong Feng, Xingyu Shen, Haowei Du, Dongyan Zhao

We convert KB subgraphs into passages to narrow the gap between KB schemas and questions, which enables our model to benefit from recent advances in multilingual pre-trained language models (MPLMs) and cross-lingual machine reading comprehension (xMRC).

Cross-Lingual Question Answering Machine Reading Comprehension

EZInterviewer: To Improve Job Interview Performance with Mock Interview Generator

no code implementations3 Jan 2023 Mingzhe Li, Xiuying Chen, Weiheng Liao, Yang song, Tao Zhang, Dongyan Zhao, Rui Yan

The key idea is to reduce the number of parameters that rely on interview dialogs by disentangling the knowledge selector and dialog generator so that most parameters can be trained with ungrounded dialogs as well as the resume data that are not low-resource.

Follow the Timeline! Generating Abstractive and Extractive Timeline Summary in Chronological Order

1 code implementation2 Jan 2023 Xiuying Chen, Mingzhe Li, Shen Gao, Zhangming Chan, Dongyan Zhao, Xin Gao, Xiangliang Zhang, Rui Yan

Nowadays, time-stamped web documents related to a general news query floods spread throughout the Internet, and timeline summarization targets concisely summarizing the evolution trajectory of events along the timeline.

Document Summarization Timeline Summarization +1

Adam: Dense Retrieval Distillation with Adaptive Dark Examples

no code implementations20 Dec 2022 Chang Liu, Chongyang Tao, Xiubo Geng, Tao Shen, Dongyan Zhao, Can Xu, Binxing Jiao, Daxin Jiang

Different from previous works that only rely on one positive and hard negatives as candidate passages, we create dark examples that all have moderate relevance to the query through mixing-up and masking in discrete space.

Knowledge Distillation Retrieval

Neural Machine Translation with Contrastive Translation Memories

1 code implementation6 Dec 2022 Xin Cheng, Shen Gao, Lemao Liu, Dongyan Zhao, Rui Yan

Retrieval-augmented Neural Machine Translation models have been successful in many translation scenarios.

Contrastive Learning Machine Translation +4

Do Charge Prediction Models Learn Legal Theory?

1 code implementation31 Oct 2022 Zhenwei An, Quzhe Huang, Cong Jiang, Yansong Feng, Dongyan Zhao

The charge prediction task aims to predict the charge for a case given its fact description.

Towards Efficient Dialogue Pre-training with Transferable and Interpretable Latent Structure

no code implementations22 Oct 2022 Xueliang Zhao, Lemao Liu, Tingchen Fu, Shuming Shi, Dongyan Zhao, Rui Yan

With the availability of massive general-domain dialogue data, pre-trained dialogue generation appears to be super appealing to transfer knowledge from the general domain to downstream applications.

Dialogue Generation

Counterfactual Recipe Generation: Exploring Compositional Generalization in a Realistic Scenario

1 code implementation20 Oct 2022 Xiao Liu, Yansong Feng, Jizhi Tang, Chengang Hu, Dongyan Zhao

Although pretrained language models can generate fluent recipe texts, they fail to truly learn and use the culinary knowledge in a compositional way.

counterfactual Recipe Generation

Knowledge-enhanced Iterative Instruction Generation and Reasoning for Knowledge Base Question Answering

no code implementations7 Sep 2022 Haowei Du, Quzhe Huang, Chen Zhang, Dongyan Zhao

Multi-hop Knowledge Base Question Answering(KBQA) aims to find the answer entity in a knowledge base which is several hops from the topic entity mentioned in the question.

Knowledge Base Question Answering Retrieval

GNN-encoder: Learning a Dual-encoder Architecture via Graph Neural Networks for Dense Passage Retrieval

no code implementations18 Apr 2022 Jiduan Liu, Jiahao Liu, Yang Yang, Jingang Wang, Wei Wu, Dongyan Zhao, Rui Yan

To enhance the performance of dense retrieval models without loss of efficiency, we propose a GNN-encoder model in which query (passage) information is fused into passage (query) representations via graph neural networks that are constructed by queries and their top retrieved passages.

Natural Questions Passage Retrieval +2

Learning to Express in Knowledge-Grounded Conversation

no code implementations NAACL 2022 Xueliang Zhao, Tingchen Fu, Chongyang Tao, Wei Wu, Dongyan Zhao, Rui Yan

Grounding dialogue generation by extra knowledge has shown great potentials towards building a system capable of replying with knowledgeable and engaging responses.

Dialogue Generation

Entailment Graph Learning with Textual Entailment and Soft Transitivity

1 code implementation ACL 2022 Zhibin Chen, Yansong Feng, Dongyan Zhao

Typed entailment graphs try to learn the entailment relations between predicates from text and model them as edges between predicate nodes.

Graph Learning Natural Language Inference

Things not Written in Text: Exploring Spatial Commonsense from Visual Signals

1 code implementation ACL 2022 Xiao Liu, Da Yin, Yansong Feng, Dongyan Zhao

We probe PLMs and models with visual signals, including vision-language pretrained models and image synthesis models, on this benchmark, and find that image synthesis models are more capable of learning accurate and consistent spatial knowledge than other models.

Image Generation Natural Language Understanding +1

Stylized Dialogue Generation with Multi-Pass Dual Learning

1 code implementation NeurIPS 2021 Jinpeng Li, Yingce Xia, Rui Yan, Hongda Sun, Dongyan Zhao, Tie-Yan Liu

Considering there is no parallel data between the contexts and the responses of target style S1, existing works mainly use back translation to generate stylized synthetic data for training, where the data about context, target style S1 and an intermediate style S0 is used.

Dialogue Generation

Extract, Integrate, Compete: Towards Verification Style Reading Comprehension

1 code implementation Findings (EMNLP) 2021 Chen Zhang, Yuxuan Lai, Yansong Feng, Dongyan Zhao

In this paper, we present a new verification style reading comprehension dataset named VGaokao from Chinese Language tests of Gaokao.

Reading Comprehension

Response Ranking with Multi-types of Deep Interactive Representations in Retrieval-based Dialogues

1 code implementation ACM Transactions on Information Systems 2021 Ruijian Xu, Chongyang Tao, Jiazhan Feng, Wei Wu, Rui Yan, Dongyan Zhao

To tackle these challenges, we propose a representation[K]-interaction[L]-matching framework that explores multiple types of deep interactive representations to build context-response matching models for response selection.

Conversational Response Selection Retrieval

Capturing Relations between Scientific Papers: An Abstractive Model for Related Work Section Generation

1 code implementation ACL 2021 Xiuying Chen, Hind Alamro, Mingzhe Li, Shen Gao, Xiangliang Zhang, Dongyan Zhao, Rui Yan

Hence, in this paper, we propose a Relation-aware Related work Generator (RRG), which generates an abstractive related work from the given multiple scientific papers in the same research area.

Relation

Exploring Distantly-Labeled Rationales in Neural Network Models

no code implementations ACL 2021 Quzhe Huang, Shengqi Zhu, Yansong Feng, Dongyan Zhao

Recent studies strive to incorporate various human rationales into neural networks to improve model performance, but few pay attention to the quality of the rationales.

Why Machine Reading Comprehension Models Learn Shortcuts?

1 code implementation Findings (ACL) 2021 Yuxuan Lai, Chen Zhang, Yansong Feng, Quzhe Huang, Dongyan Zhao

A thorough empirical analysis shows that MRC models tend to learn shortcut questions earlier than challenging questions, and the high proportions of shortcut questions in training sets hinder models from exploring the sophisticated reasoning skills in the later stage of training.

Machine Reading Comprehension

Learning to Organize a Bag of Words into Sentences with Neural Networks: An Empirical Study

no code implementations NAACL 2021 Chongyang Tao, Shen Gao, Juntao Li, Yansong Feng, Dongyan Zhao, Rui Yan

Sequential information, a. k. a., orders, is assumed to be essential for processing a sequence with recurrent neural network or convolutional neural network based encoders.

Sentence

Lattice-BERT: Leveraging Multi-Granularity Representations in Chinese Pre-trained Language Models

2 code implementations NAACL 2021 Yuxuan Lai, Yijia Liu, Yansong Feng, Songfang Huang, Dongyan Zhao

Further analysis shows that Lattice-BERT can harness the lattice structures, and the improvement comes from the exploration of redundant information and multi-granularity representations.

Natural Language Understanding Sentence

Dialogue History Matters! Personalized Response Selectionin Multi-turn Retrieval-based Chatbots

no code implementations17 Mar 2021 Juntao Li, Chang Liu, Chongyang Tao, Zhangming Chan, Dongyan Zhao, Min Zhang, Rui Yan

To fill the gap between these up-to-date methods and the real-world applications, we incorporate user-specific dialogue history into the response selection and propose a personalized hybrid matching network (PHMN).

Representation Learning Retrieval

The Style-Content Duality of Attractiveness: Learning to Write Eye-Catching Headlines via Disentanglement

no code implementations14 Dec 2020 Mingzhe Li, Xiuying Chen, Min Yang, Shen Gao, Dongyan Zhao, Rui Yan

In this paper, we propose a Disentanglement-based Attractive Headline Generator (DAHG) that generates headline which captures the attractive content following the attractive style.

Disentanglement

Reasoning in Dialog: Improving Response Generation by Context Reading Comprehension

1 code implementation14 Dec 2020 Xiuying Chen, Zhi Cui, Jiayi Zhang, Chen Wei, Jianwei Cui, Bin Wang, Dongyan Zhao, Rui Yan

Hence, in this paper, we propose to improve the response generation performance by examining the model's ability to answer a reading comprehension question, where the question is focused on the omitted information in the dialog.

Multi-Task Learning Reading Comprehension +1

Meaningful Answer Generation of E-Commerce Question-Answering

no code implementations14 Nov 2020 Shen Gao, Xiuying Chen, Zhaochun Ren, Dongyan Zhao, Rui Yan

To generate more meaningful answers, in this paper, we propose a novel generative neural model, called the Meaningful Product Answer Generator (MPAG), which alleviates the safe answer problem by taking product reviews, product attributes, and a prototype answer into consideration.

Answer Generation Question Answering +1

Learning to Respond with Your Favorite Stickers: A Framework of Unifying Multi-Modality and User Preference in Multi-Turn Dialog

no code implementations5 Nov 2020 Shen Gao, Xiuying Chen, Li Liu, Dongyan Zhao, Rui Yan

Hence, in this paper, we propose to recommend an appropriate sticker to user based on multi-turn dialog context and sticker using history of user.

Towards Context-Aware Code Comment Generation

no code implementations Findings of the Association for Computational Linguistics 2020 Xiaohan Yu, Quzhe Huang, Zheng Wang, Yansong Feng, Dongyan Zhao

Code comments are vital for software maintenance and comprehension, but many software projects suffer from the lack of meaningful and up-to-date comments in practice.

Code Comment Generation Comment Generation +1

VMSMO: Learning to Generate Multimodal Summary for Video-based News Articles

1 code implementation EMNLP 2020 Mingzhe Li, Xiuying Chen, Shen Gao, Zhangming Chan, Dongyan Zhao, Rui Yan

Hence, in this paper, we propose the task of Video-based Multimodal Summarization with Multimodal Output (VMSMO) to tackle such a problem.

Learning an Effective Context-Response Matching Model with Self-Supervised Tasks for Retrieval-based Dialogues

no code implementations14 Sep 2020 Ruijian Xu, Chongyang Tao, Daxin Jiang, Xueliang Zhao, Dongyan Zhao, Rui Yan

To address these issues, in this paper, we propose learning a context-response matching model with auxiliary self-supervised tasks designed for the dialogue data based on pre-trained language models.

Conversational Response Selection Retrieval

Domain Adaptation for Semantic Parsing

no code implementations23 Jun 2020 Zechang Li, Yuxuan Lai, Yansong Feng, Dongyan Zhao

In this paper, we propose a novel semantic parser for domain adaptation, where we have much fewer annotated data in the target domain compared to the source domain.

Domain Adaptation Semantic Parsing

Cross-Lingual Low-Resource Set-to-Description Retrieval for Global E-Commerce

1 code implementation17 May 2020 Juntao Li, Chang Liu, Jian Wang, Lidong Bing, Hongsong Li, Xiaozhong Liu, Dongyan Zhao, Rui Yan

We manually collect a new and high-quality paired dataset, where each pair contains an unordered product attribute set in the source language and an informative product description in the target language.

Attribute Cross-Lingual Information Retrieval +1

Neighborhood Matching Network for Entity Alignment

1 code implementation ACL 2020 Yuting Wu, Xiao Liu, Yansong Feng, Zheng Wang, Dongyan Zhao

This paper presents Neighborhood Matching Network (NMN), a novel entity alignment framework for tackling the structural heterogeneity challenge.

Entity Alignment Graph Sampling +1

From Standard Summarization to New Tasks and Beyond: Summarization with Manifold Information

no code implementations10 May 2020 Shen Gao, Xiuying Chen, Zhaochun Ren, Dongyan Zhao, Rui Yan

Text summarization is the research area aiming at creating a short and condensed version of the original document, which conveys the main idea of the document in a few words.

Text Summarization

Learning to Respond with Stickers: A Framework of Unifying Multi-Modality in Multi-Turn Dialog

1 code implementation10 Mar 2020 Shen Gao, Xiuying Chen, Chang Liu, Li Liu, Dongyan Zhao, Rui Yan

Stickers with vivid and engaging expressions are becoming increasingly popular in online messaging apps, and some works are dedicated to automatically select sticker response by matching text labels of stickers with previous utterances.

Low-Resource Knowledge-Grounded Dialogue Generation

no code implementations ICLR 2020 Xueliang Zhao, Wei Wu, Chongyang Tao, Can Xu, Dongyan Zhao, Rui Yan

In such a low-resource setting, we devise a disentangled response decoder in order to isolate parameters that depend on knowledge-grounded dialogues from the entire generation model.

Dialogue Generation Response Generation

Integrating Relation Constraints with Neural Relation Extractors

1 code implementation26 Nov 2019 Yuan Ye, Yansong Feng, Bingfeng Luo, Yuxuan Lai, Dongyan Zhao

However, such models often make predictions for each entity pair individually, thus often fail to solve the inconsistency among different predictions, which can be characterized by discrete relation constraints.

Relation Relation Extraction

Query-bag Matching with Mutual Coverage for Information-seeking Conversations in E-commerce

1 code implementation7 Nov 2019 Zhenxin Fu, Feng Ji, Wenpeng Hu, Wei Zhou, Dongyan Zhao, Haiqing Chen, Rui Yan

Information-seeking conversation system aims at satisfying the information needs of users through conversations.

Text Matching

Learning to Update Knowledge Graphs by Reading News

no code implementations IJCNLP 2019 Jizhi Tang, Yansong Feng, Dongyan Zhao

News streams contain rich up-to-date information which can be used to update knowledge graphs (KGs).

Knowledge Graphs

Who Is Speaking to Whom? Learning to Identify Utterance Addressee in Multi-Party Conversations

no code implementations IJCNLP 2019 Ran Le, Wenpeng Hu, Mingyue Shang, Zhenjun You, Lidong Bing, Dongyan Zhao, Rui Yan

Previous research on dialogue systems generally focuses on the conversation between two participants, yet multi-party conversations which involve more than two participants within one session bring up a more complicated but realistic scenario.

Stick to the Facts: Learning towards a Fidelity-oriented E-Commerce Product Description Generation

no code implementations IJCNLP 2019 Zhangming Chan, Xiuying Chen, Yongliang Wang, Juntao Li, Zhiqiang Zhang, Kun Gai, Dongyan Zhao, Rui Yan

Different from other text generation tasks, in product description generation, it is of vital importance to generate faithful descriptions that stick to the product attribute information.

Attribute Text Generation

RPM-Oriented Query Rewriting Framework for E-commerce Keyword-Based Sponsored Search

no code implementations28 Oct 2019 Xiuying Chen, Daorui Xiao, Shen Gao, Guojun Liu, Wei. Lin, Bo Zheng, Dongyan Zhao, Rui Yan

Sponsored search optimizes revenue and relevance, which is estimated by Revenue Per Mille (RPM).

Multilingual Dialogue Generation with Shared-Private Memory

no code implementations6 Oct 2019 Chen Chen, Lisong Qiu, Zhenxin Fu, Dongyan Zhao, Junfei Liu, Rui Yan

Existing dialog systems are all monolingual, where features shared among different languages are rarely explored.

Cross-Lingual Transfer Dialogue Generation

Semi-supervised Text Style Transfer: Cross Projection in Latent Space

no code implementations IJCNLP 2019 Mingyue Shang, Piji Li, Zhenxin Fu, Lidong Bing, Dongyan Zhao, Shuming Shi, Rui Yan

Text style transfer task requires the model to transfer a sentence of one style to another style while retaining its original content meaning, which is a challenging problem that has long suffered from the shortage of parallel data.

Sentence Style Transfer +1

Learning from Positive and Unlabeled Data with Adversarial Training

no code implementations25 Sep 2019 Wenpeng Hu, Ran Le, Bing Liu, Feng Ji, Haiqing Chen, Dongyan Zhao, Jinwen Ma, Rui Yan

Positive-unlabeled (PU) learning learns a binary classifier using only positive and unlabeled examples without labeled negative examples.

Jointly Learning Entity and Relation Representations for Entity Alignment

1 code implementation IJCNLP 2019 Yuting Wu, Xiao Liu, Yansong Feng, Zheng Wang, Dongyan Zhao

Entity alignment is a viable means for integrating heterogeneous knowledge among different knowledge graphs (KGs).

Ranked #18 on Entity Alignment on DBP15k zh-en (using extra training data)

Entity Alignment Entity Embeddings +2

How to Write Summaries with Patterns? Learning towards Abstractive Summarization through Prototype Editing

1 code implementation IJCNLP 2019 Shen Gao, Xiuying Chen, Piji Li, Zhangming Chan, Dongyan Zhao, Rui Yan

There are two main challenges in this task: (1) the model needs to incorporate learned patterns from the prototype, but (2) should avoid copying contents other than the patternized words---such as irrelevant facts---into the generated summaries.

Abstractive Text Summarization

A Sketch-Based System for Semantic Parsing

1 code implementation2 Sep 2019 Zechang Li, Yuxuan Lai, Yuxi Xie, Yansong Feng, Dongyan Zhao

The sketch is a high-level structure of the logical form exclusive of low-level details such as entities and predicates.

Semantic Parsing Task 2

Relation-Aware Entity Alignment for Heterogeneous Knowledge Graphs

1 code implementation22 Aug 2019 Yuting Wu, Xiao Liu, Yansong Feng, Zheng Wang, Rui Yan, Dongyan Zhao

Entity alignment is the task of linking entities with the same real-world identity from different knowledge graphs (KGs), which has been recently dominated by embedding-based methods.

Ranked #20 on Entity Alignment on DBP15k zh-en (using extra training data)

Entity Alignment Entity Embeddings +2

Learning towards Abstractive Timeline Summarization

1 code implementation IJCAI 2019 2019 Xiuying Chen, Zhangming Chan, Shen Gao, Meng-Hsuan Yu, Dongyan Zhao, Rui Yan

Timeline summarization targets at concisely summarizing the evolution trajectory along the timeline and existing timeline summarization approaches are all based on extractive methods. In this paper, we propose the task of abstractive timeline summarization, which tends to concisely paraphrase the information in the time-stamped events. Unlike traditional document summarization, timeline summarization needs to model the time series information of the input events and summarize important events in chronological order. To tackle this challenge, we propose a memory-based timeline summarization model (MTS). Concretely, we propose a time-event memory to establish a timeline, and use the time position of events on this timeline to guide generation process. Besides, in each decoding step, we incorporate event-level information into word-level attention to avoid confusion between events. Extensive experiments are conducted on a large-scale real-world dataset, and the results show that MTS achieves the state-of-the-art performance in terms of both automatic and human evaluations.

Document Summarization Timeline Summarization +2

Are Training Samples Correlated? Learning to Generate Dialogue Responses with Multiple References

no code implementations ACL 2019 Lisong Qiu, Juntao Li, Wei Bi, Dongyan Zhao, Rui Yan

Due to its potential applications, open-domain dialogue generation has become popular and achieved remarkable progress in recent years, but sometimes suffers from generic responses.

Dialogue Generation valid

A Document-grounded Matching Network for Response Selection in Retrieval-based Chatbots

no code implementations11 Jun 2019 Xueliang Zhao, Chongyang Tao, Wei Wu, Can Xu, Dongyan Zhao, Rui Yan

We present a document-grounded matching network (DGMN) for response selection that can power a knowledge-aware retrieval-based chatbot system.

Chatbot Retrieval

GSN: A Graph-Structured Network for Multi-Party Dialogues

1 code implementation31 May 2019 Wenpeng Hu, Zhangming Chan, Bing Liu, Dongyan Zhao, Jinwen Ma, Rui Yan

Existing neural models for dialogue response generation assume that utterances are sequentially organized.

Response Generation

Lattice CNNs for Matching Based Chinese Question Answering

1 code implementation25 Feb 2019 Yuxuan Lai, Yansong Feng, Xiaohan Yu, Zheng Wang, Kun Xu, Dongyan Zhao

Short text matching often faces the challenges that there are great word mismatch and expression diversity between the two texts, which would be further aggravated in languages like Chinese where there is no natural space to segment words explicitly.

Question Answering Text Matching

Product-Aware Answer Generation in E-Commerce Question-Answering

1 code implementation23 Jan 2019 Shen Gao, Zhaochun Ren, Yihong Eric Zhao, Dongyan Zhao, Dawei Yin, Rui Yan

In this paper, we propose the task of product-aware answer generation, which tends to generate an accurate and complete answer from large-scale unlabeled e-commerce reviews and product attributes.

Answer Generation Question Answering

Find a Reasonable Ending for Stories: Does Logic Relation Help the Story Cloze Test?

no code implementations13 Dec 2018 Mingyue Shang, Zhenxin Fu, Hongzhi Yin, Bo Tang, Dongyan Zhao, Rui Yan

In this paper, we incorporate the logic information with the help of the Natural Language Inference (NLI) task to the Story Cloze Test (SCT).

Cloze Test Natural Language Inference +2

Abstractive Text Summarization by Incorporating Reader Comments

no code implementations13 Dec 2018 Shen Gao, Xiuying Chen, Piji Li, Zhaochun Ren, Lidong Bing, Dongyan Zhao, Rui Yan

To tackle this problem, we propose the task of reader-aware abstractive summary generation, which utilizes the reader comments to help the model produce better summary about the main aspect.

Reader-Aware Summarization

Chat More If You Like: Dynamic Cue Words Planning to Flow Longer Conversations

no code implementations19 Nov 2018 Lili Yao, Ruijian Xu, Chao Li, Dongyan Zhao, Rui Yan

To build an open-domain multi-turn conversation system is one of the most interesting and challenging tasks in Artificial Intelligence.

Plan-And-Write: Towards Better Automatic Storytelling

2 code implementations14 Nov 2018 Lili Yao, Nanyun Peng, Ralph Weischedel, Kevin Knight, Dongyan Zhao, Rui Yan

Automatic storytelling is challenging since it requires generating long, coherent natural language to describes a sensible sequence of events.

Story Generation

Encoding Implicit Relation Requirements for Relation Extraction: A Joint Inference Approach

no code implementations9 Nov 2018 Li-Wei Chen, Yansong Feng, Songfang Huang, Bingfeng Luo, Dongyan Zhao

Relation extraction is the task of identifying predefined relationship between entities, and plays an essential role in information extraction, knowledge base construction, question answering and so on.

Question Answering Relation +1

Iterative Document Representation Learning Towards Summarization with Polishing

1 code implementation EMNLP 2018 Xiuying Chen, Shen Gao, Chongyang Tao, Yan Song, Dongyan Zhao, Rui Yan

In this paper, we introduce Iterative Text Summarization (ITS), an iteration-based model for supervised extractive text summarization, inspired by the observation that it is often necessary for a human to read an article multiple times in order to fully understand and summarize its contents.

Extractive Text Summarization Representation Learning +1

Improving Matching Models with Hierarchical Contextualized Representations for Multi-turn Response Selection

no code implementations22 Aug 2018 Chongyang Tao, Wei Wu, Can Xu, Yansong Feng, Dongyan Zhao, Rui Yan

In this paper, we study context-response matching with pre-trained contextualized representations for multi-turn response selection in retrieval-based chatbots.

Dialogue Generation Retrieval +1

Marrying up Regular Expressions with Neural Networks: A Case Study for Spoken Language Understanding

no code implementations ACL 2018 Bingfeng Luo, Yansong Feng, Zheng Wang, Songfang Huang, Rui Yan, Dongyan Zhao

The success of many natural language processing (NLP) tasks is bound by the number and quality of annotated data, but there is often a shortage of such training data.

Intent Detection slot-filling +2

Topic-Based Question Generation

no code implementations ICLR 2018 Wenpeng Hu, Bing Liu, Rui Yan, Dongyan Zhao, Jinwen Ma

In the paper, we propose a new question generation problem, which also requires the input of a target topic in addition to a piece of descriptive text.

Chatbot Descriptive +3

Tree2Tree Learning with Memory Unit

no code implementations ICLR 2018 Ning Miao, Hengliang Wang, Ran Le, Chongyang Tao, Mingyue Shang, Rui Yan, Dongyan Zhao

Traditional recurrent neural network (RNN) or convolutional neural net- work (CNN) based sequence-to-sequence model can not handle tree structural data well.

Machine Translation Translation

Scale Up Event Extraction Learning via Automatic Training Data Generation

no code implementations11 Dec 2017 Ying Zeng, Yansong Feng, Rong Ma, Zheng Wang, Rui Yan, Chongde Shi, Dongyan Zhao

We show that this large volume of training data not only leads to a better event extractor, but also allows us to detect multiple typed events.

Event Extraction

Style Transfer in Text: Exploration and Evaluation

2 code implementations18 Nov 2017 Zhenxin Fu, Xiaoye Tan, Nanyun Peng, Dongyan Zhao, Rui Yan

Results show that the proposed content preservation metric is highly correlate to human judgments, and the proposed models are able to generate sentences with higher style transfer strength and similar content preservation score comparing to auto-encoder.

Style Transfer Text Style Transfer

Diversifying Neural Conversation Model with Maximal Marginal Relevance

no code implementations IJCNLP 2017 Yiping Song, Zhiliang Tian, Dongyan Zhao, Ming Zhang, Rui Yan

However, traditional seq2seq suffer from a severe weakness: during beam search decoding, they tend to rank universal replies at the top of the candidate list, resulting in the lack of diversity among candidate replies.

Document Summarization Information Retrieval +1

Learning to Predict Charges for Criminal Cases with Legal Basis

no code implementations EMNLP 2017 Bingfeng Luo, Yansong Feng, Jianbo Xu, Xiang Zhang, Dongyan Zhao

The charge prediction task is to determine appropriate charges for a given case, which is helpful for legal assistant systems where the user input is fact description.

A Constrained Sequence-to-Sequence Neural Model for Sentence Simplification

no code implementations7 Apr 2017 Yaoyuan Zhang, Zhenxu Ye, Yansong Feng, Dongyan Zhao, Rui Yan

For word-level studies, words are simplified but also have potential grammar errors due to different usages of words before and after simplification.

Sentence

RUBER: An Unsupervised Method for Automatic Evaluation of Open-Domain Dialog Systems

1 code implementation11 Jan 2017 Chongyang Tao, Lili Mou, Dongyan Zhao, Rui Yan

Open-domain human-computer conversation has been attracting increasing attention over the past few years.

Dialogue Evaluation Retrieval

Hybrid Question Answering over Knowledge Base and Free Text

no code implementations COLING 2016 Kun Xu, Yansong Feng, Songfang Huang, Dongyan Zhao

While these systems are able to provide more precise answers than information retrieval (IR) based QA systems, the natural incompleteness of KB inevitably limits the question scope that the system can answer.

Information Retrieval Question Answering +2

Two are Better than One: An Ensemble of Retrieval- and Generation-Based Dialog Systems

2 code implementations23 Oct 2016 Yiping Song, Rui Yan, Xiang Li, Dongyan Zhao, Ming Zhang

In this paper, we propose a novel ensemble of retrieval-based and generation-based dialog systems in the open domain.

Retrieval

Cannot find the paper you are looking for? You can Submit a new open access paper.