1 code implementation • 27 Mar 2024 • Shawn Im, Yixuan Li
Our work provides an initial attempt to theoretically analyze the learning dynamics of human preference alignment.
no code implementations • 10 Dec 2023 • Shawn Im, Jacob Andreas, Yilun Zhou
One of the motivations for explainable AI is to allow humans to make better and more informed decisions regarding the use and deployment of AI models.