1 code implementation • 11 Jan 2024 • Asma Ghandeharioun, Avi Caciularu, Adam Pearce, Lucas Dixon, Mor Geva
Inspecting the information encoded in hidden representations of large language models (LLMs) can explain models' behavior and verify their alignment with human values.
no code implementations • 17 Nov 2021 • Thomas McGrath, Andrei Kapishnikov, Nenad Tomašev, Adam Pearce, Demis Hassabis, Been Kim, Ulrich Paquet, Vladimir Kramnik
In this work we provide evidence that human knowledge is acquired by the AlphaZero neural network as it trains on the game of chess.
no code implementations • 14 Apr 2021 • Tolga Bolukbasi, Adam Pearce, Ann Yuan, Andy Coenen, Emily Reif, Fernanda Viégas, Martin Wattenberg
We describe an "interpretability illusion" that arises when analyzing the BERT model.
no code implementations • 21 Apr 2020 • Gregory A. Wellenius, Swapnil Vispute, Valeria Espinosa, Alex Fabrikant, Thomas C. Tsai, Jonathan Hennessy, Andrew Dai, Brian Williams, Krishna Gadepalli, Adam Boulanger, Adam Pearce, Chaitanya Kamath, Arran Schlosberg, Catherine Bendebury, Chinmoy Mandayam, Charlotte Stanton, Shailesh Bavadekar, Christopher Pluntke, Damien Desfontaines, Benjamin Jacobson, Zan Armstrong, Bryant Gipson, Royce Wilson, Andrew Widdowson, Katherine Chou, Andrew Oplinger, Tomer Shekel, Ashish K. Jha, Evgeniy Gabrilovich
Social distancing remains an important strategy to combat the COVID-19 pandemic in the United States.
2 code implementations • NeurIPS 2019 • Andy Coenen, Emily Reif, Ann Yuan, Been Kim, Adam Pearce, Fernanda Viégas, Martin Wattenberg
Transformer architectures show significant promise for natural language processing.