Search Results for author: Jiale Cheng

Found 8 papers, 7 papers with code

CritiqueLLM: Scaling LLM-as-Critic for Effective and Explainable Evaluation of Large Language Model Generation

2 code implementations30 Nov 2023 Pei Ke, Bosi Wen, Zhuoer Feng, Xiao Liu, Xuanyu Lei, Jiale Cheng, Shengyuan Wang, Aohan Zeng, Yuxiao Dong, Hongning Wang, Jie Tang, Minlie Huang

Since the natural language processing (NLP) community started to make large language models (LLMs), such as GPT-4, act as a critic to evaluate the quality of generated texts, most of them only train a critique generation model of a specific scale on specific datasets.

Language Modelling Large Language Model

Black-Box Prompt Optimization: Aligning Large Language Models without Model Training

1 code implementation7 Nov 2023 Jiale Cheng, Xiao Liu, Kehan Zheng, Pei Ke, Hongning Wang, Yuxiao Dong, Jie Tang, Minlie Huang

However, these models are often not well aligned with human intents, which calls for additional treatments on them, that is, the alignment problem.

Safety Assessment of Chinese Large Language Models

2 code implementations20 Apr 2023 Hao Sun, Zhexin Zhang, Jiawen Deng, Jiale Cheng, Minlie Huang

To further promote the safe deployment of LLMs, we develop a Chinese LLM safety assessment benchmark.

Towards Safer Generative Language Models: A Survey on Safety Risks, Evaluations, and Improvements

no code implementations18 Feb 2023 Jiawen Deng, Jiale Cheng, Hao Sun, Zhexin Zhang, Minlie Huang

This survey presents a framework for safety research pertaining to large models, delineating the landscape of safety risks as well as safety evaluation and improvement methods.

Adversarial Attack Ethics

PAL: Persona-Augmented Emotional Support Conversation Generation

1 code implementation19 Dec 2022 Jiale Cheng, Sahand Sabour, Hao Sun, Zhuang Chen, Minlie Huang

As previous studies have demonstrated that seekers' persona is an important factor for effective support, we investigate whether there are benefits to modeling such information in dialogue models for support.

Constructing Highly Inductive Contexts for Dialogue Safety through Controllable Reverse Generation

1 code implementation4 Dec 2022 Zhexin Zhang, Jiale Cheng, Hao Sun, Jiawen Deng, Fei Mi, Yasheng Wang, Lifeng Shang, Minlie Huang

In order to detect such toxic generations, existing methods rely on templates, real-world data extraction, crowdsourcing workers, or automatic generation to construct adversarial contexts that are likely to induce toxic generations.

Response Generation

On the Safety of Conversational Models: Taxonomy, Dataset, and Benchmark

1 code implementation Findings (ACL) 2022 Hao Sun, Guangxuan Xu, Jiawen Deng, Jiale Cheng, Chujie Zheng, Hao Zhou, Nanyun Peng, Xiaoyan Zhu, Minlie Huang

We propose a taxonomy for dialogue safety specifically designed to capture unsafe behaviors in human-bot dialogue settings, with focuses on context-sensitive unsafety, which is under-explored in prior works.

Cannot find the paper you are looking for? You can Submit a new open access paper.