Search Results for author: Shawn Im

Found 2 papers, 1 papers with code

Understanding the Learning Dynamics of Alignment with Human Feedback

1 code implementation27 Mar 2024 Shawn Im, Yixuan Li

Our work provides an initial attempt to theoretically analyze the learning dynamics of human preference alignment.

Evaluating the Utility of Model Explanations for Model Development

no code implementations10 Dec 2023 Shawn Im, Jacob Andreas, Yilun Zhou

One of the motivations for explainable AI is to allow humans to make better and more informed decisions regarding the use and deployment of AI models.

counterfactual Decision Making +1

Cannot find the paper you are looking for? You can Submit a new open access paper.