Search Results for author: Maxim Khanov

Found 1 papers, 1 papers with code

ARGS: Alignment as Reward-Guided Search

1 code implementation23 Jan 2024 Maxim Khanov, Jirayu Burapacheep, Yixuan Li

Aligning large language models with human objectives is paramount, yet common approaches including RLHF suffer from unstable and resource-intensive training.

Cannot find the paper you are looking for? You can Submit a new open access paper.