no code implementations • NeurIPS 2023 • Fulton Wang, Julius Adebayo, Sarah Tan, Diego Garcia-Olano, Narine Kokhlikyan
We present a method for identifying groups of test examples -- slices -- on which a model under-performs, a task now known as slice discovery.
no code implementations • 23 Apr 2023 • Zhi Chen, Sarah Tan, Urszula Chajewska, Cynthia Rudin, Rich Caruana
Missing values are a fundamental problem in data science.
no code implementations • 30 Mar 2023 • Mia Garrard, Hanson Wang, Ben Letham, Shaun Singh, Abbas Kazerouni, Sarah Tan, Zehui Wang, Yin Huang, Yichun Hu, Chad Zhou, Norm Zhou, Eytan Bakshy
Many organizations measure treatment effects via an experimentation platform to evaluate the casual effect of product variations prior to full-scale deployment.
no code implementations • 10 Jun 2022 • Leon Yao, Caroline Lo, Israel Nir, Sarah Tan, Ariel Evnine, Adam Lerer, Alex Peysakhovich
Learning heterogeneous treatment effects (HTEs) is an important problem across many fields.
no code implementations • 5 Nov 2021 • Han Wu, Sarah Tan, Weiwei Li, Mia Garrard, Adam Obeng, Drew Dimmery, Shaun Singh, Hanson Wang, Daniel Jiang, Eytan Bakshy
Black-box heterogeneous treatment effect (HTE) models are increasingly being used to create personalized policies that assign individuals to their optimal treatments.
2 code implementations • 11 Jun 2020 • Chun-Hao Chang, Sarah Tan, Ben Lengerich, Anna Goldenberg, Rich Caruana
Generalized additive models (GAMs) have become a leading modelclass for interpretable machine learning.
1 code implementation • 12 Nov 2019 • Benjamin Lengerich, Sarah Tan, Chun-Hao Chang, Giles Hooker, Rich Caruana
Models which estimate main effects of individual variables alongside interaction effects have an identifiability challenge: effects can be freely moved between main effects and interaction effects without changing the model prediction.
no code implementations • 29 Apr 2019 • Yu-jia Zhang, Kuangyan Song, Yiming Sun, Sarah Tan, Madeleine Udell
Methods for interpreting machine learning black-box models increase the outcomes' transparency and in turn generates insight into the reliability and fairness of the algorithms.
1 code implementation • 22 Oct 2018 • Xuezhou Zhang, Sarah Tan, Paul Koch, Yin Lou, Urszula Chajewska, Rich Caruana
In the first part of this paper, we generalize a state-of-the-art GAM learning algorithm based on boosted trees to the multiclass setting, and show that this multiclass algorithm outperforms existing GAM learning algorithms and sometimes matches the performance of full complexity models such as gradient boosted trees.
no code implementations • 28 Aug 2018 • Sarah Tan, Julius Adebayo, Kori Inkpen, Ece Kamar
Dressel and Farid (2018) asked Mechanical Turk workers to evaluate a subset of defendants in the ProPublica COMPAS data for risk of recidivism, and concluded that COMPAS predictions were no more accurate or fair than predictions made by humans.
1 code implementation • ICLR 2019 • Sarah Tan, Giles Hooker, Paul Koch, Albert Gordo, Rich Caruana
In this paper, we study global additive explanations for non-additive models, focusing on four explanation methods: partial dependence, Shapley explanations adapted to a global setting, distilled additive explanations, and gradient-based explanations.
no code implementations • 19 Nov 2017 • Skyler Seto, Sarah Tan, Giles Hooker, Martin T. Wells
Non-negative matrix factorization (NMF) is a technique for finding latent representations of data.
1 code implementation • 17 Oct 2017 • Sarah Tan, Rich Caruana, Giles Hooker, Yin Lou
We compare the student model trained with distillation to a second un-distilled transparent model trained on ground-truth outcomes, and use differences between the two models to gain insight into the black-box model.
2 code implementations • 22 Nov 2016 • Sarah Tan, Matvey Soloviev, Giles Hooker, Martin T. Wells
Ensembles of decision trees perform well on many problems, but are not interpretable.