Search Results for author: Haozhe Ji

Found 13 papers, 10 papers with code

Towards Efficient and Exact Optimization of Language Model Alignment

2 code implementations1 Feb 2024 Haozhe Ji, Cheng Lu, Yilin Niu, Pei Ke, Hongning Wang, Jun Zhu, Jie Tang, Minlie Huang

We prove that EXO is guaranteed to optimize in the same direction as the RL algorithms asymptotically for arbitary parametrization of the policy, while enables efficient optimization by circumventing the complexities associated with RL algorithms.

Language Modelling Reinforcement Learning (RL)

Language Model Decoding as Direct Metrics Optimization

no code implementations2 Oct 2023 Haozhe Ji, Pei Ke, Hongning Wang, Minlie Huang

And most importantly, we prove that this induced distribution is guaranteed to improve the perplexity on human texts, which suggests a better approximation to the underlying distribution of human texts.

Language Modelling

Tailoring Language Generation Models under Total Variation Distance

1 code implementation26 Feb 2023 Haozhe Ji, Pei Ke, Zhipeng Hu, Rongsheng Zhang, Minlie Huang

The standard paradigm of neural language generation adopts maximum likelihood estimation (MLE) as the optimizing method.

Text Generation

Curriculum-Based Self-Training Makes Better Few-Shot Learners for Data-to-Text Generation

1 code implementation6 Jun 2022 Pei Ke, Haozhe Ji, Zhenyu Yang, Yi Huang, Junlan Feng, Xiaoyan Zhu, Minlie Huang

Despite the success of text-to-text pre-trained models in various natural language generation (NLG) tasks, the generation performance is largely restricted by the number of labeled data in downstream tasks, particularly in data-to-text generation tasks.

Data-to-Text Generation Unsupervised Pre-training

LaMemo: Language Modeling with Look-Ahead Memory

1 code implementation NAACL 2022 Haozhe Ji, Rongsheng Zhang, Zhenyu Yang, Zhipeng Hu, Minlie Huang

Although Transformers with fully connected self-attentions are powerful to model long-term dependencies, they are struggling to scale to long texts with thousands of words in language modeling.

Language Modelling

DiscoDVT: Generating Long Text with Discourse-Aware Discrete Variational Transformer

1 code implementation EMNLP 2021 Haozhe Ji, Minlie Huang

Despite the recent advances in applying pre-trained language models to generate high-quality texts, generating long passages that maintain long-range coherence is yet challenging for these models.

Story Generation

JointGT: Graph-Text Joint Representation Learning for Text Generation from Knowledge Graphs

1 code implementation Findings (ACL) 2021 Pei Ke, Haozhe Ji, Yu Ran, Xin Cui, LiWei Wang, Linfeng Song, Xiaoyan Zhu, Minlie Huang

Existing pre-trained models for knowledge-graph-to-text (KG-to-text) generation simply fine-tune text-to-text pre-trained models such as BART or T5 on KG-to-text datasets, which largely ignore the graph structure during encoding and lack elaborate pre-training tasks to explicitly model graph-text alignments.

Graph Reconstruction KG-to-Text Generation +3

Generating Commonsense Explanation by Extracting Bridge Concepts from Reasoning Paths

no code implementations Asian Chapter of the Association for Computational Linguistics 2020 Haozhe Ji, Pei Ke, Shaohan Huang, Furu Wei, Minlie Huang

Commonsense explanation generation aims to empower the machine's sense-making capability by generating plausible explanations to statements against commonsense.

Explanation Generation

Language Generation with Multi-Hop Reasoning on Commonsense Knowledge Graph

1 code implementation EMNLP 2020 Haozhe Ji, Pei Ke, Shaohan Huang, Furu Wei, Xiaoyan Zhu, Minlie Huang

Despite the success of generative pre-trained language models on a series of text generation tasks, they still suffer in cases where reasoning over underlying commonsense knowledge is required during generation.

Text Generation

SentiLARE: Sentiment-Aware Language Representation Learning with Linguistic Knowledge

1 code implementation EMNLP 2020 Pei Ke, Haozhe Ji, Siyang Liu, Xiaoyan Zhu, Minlie Huang

To benefit the downstream tasks in sentiment analysis, we propose a novel language representation model called SentiLARE, which introduces word-level linguistic knowledge including part-of-speech tag and sentiment polarity (inferred from SentiWordNet) into pre-trained models.

Data Augmentation Language Modelling +3

Denoise while Aggregating: Collaborative Learning in Open-Domain Question Answering

no code implementations27 Sep 2018 Haozhe Ji, Yankai Lin, Zhiyuan Liu, Maosong Sun

The open-domain question answering (OpenQA) task aims to extract answers that match specific questions from a distantly supervised corpus.

Open-Domain Question Answering Reading Comprehension +1

Cannot find the paper you are looking for? You can Submit a new open access paper.