Search Results for author: Shengyuan Wang

Found 3 papers, 3 papers with code

CritiqueLLM: Scaling LLM-as-Critic for Effective and Explainable Evaluation of Large Language Model Generation

2 code implementations30 Nov 2023 Pei Ke, Bosi Wen, Zhuoer Feng, Xiao Liu, Xuanyu Lei, Jiale Cheng, Shengyuan Wang, Aohan Zeng, Yuxiao Dong, Hongning Wang, Jie Tang, Minlie Huang

Since the natural language processing (NLP) community started to make large language models (LLMs), such as GPT-4, act as a critic to evaluate the quality of generated texts, most of them only train a critique generation model of a specific scale on specific datasets.

Language Modelling Large Language Model

Model-based Offline Policy Optimization with Adversarial Network

1 code implementation5 Sep 2023 Junming Yang, Xingguo Chen, Shengyuan Wang, Bolei Zhang

Model-based offline reinforcement learning (RL), which builds a supervised transition model with logging dataset to avoid costly interactions with the online environment, has been a promising approach for offline policy optimization.

Offline RL Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.