no code implementations • 29 Jan 2024 • Michael Feffer, Anusha Sinha, Zachary C. Lipton, Hoda Heidari
In response to rising concerns surrounding the safety, security, and trustworthiness of Generative AI (GenAI) models, practitioners and regulators alike have pointed to AI red-teaming as a key component of their strategies for identifying and mitigating these risks.
no code implementations • 19 Nov 2023 • Nari Johnson, Hoda Heidari
Artificial Intelligence Impact Assessments ("AIIAs"), a family of tools that provide structured processes to imagine the possible impacts of a proposed AI system, have become an increasingly popular proposal to govern AI systems.
no code implementations • 6 Nov 2023 • Mateo Dulce Rubio, Siqi Zeng, Qi Wang, Didier Alvarado, Francisco Moreno, Hoda Heidari, Fei Fang
Landmines remain a threat to war-affected communities for years after conflicts have ended, partly due to the laborious nature of demining tasks.
no code implementations • 10 Oct 2023 • Michael Feffer, Nikolas Martelaro, Hoda Heidari
Prior work has established the importance of integrating AI ethics topics into computer and data sciences curricula.
no code implementations • 29 Sep 2023 • Emily Black, Rakshit Naidu, Rayid Ghani, Kit T. Rodolfa, Daniel E. Ho, Hoda Heidari
While algorithmic fairness is a thriving area of research, in practice, mitigating issues of bias often gets reduced to enforcing an arbitrarily chosen fairness metric, either by enforcing fairness constraints during the optimization step, post-processing model outputs, or by manipulating the training data.
no code implementations • 8 Aug 2023 • Benjamin Laufer, Jon Kleinberg, Hoda Heidari
Both entities are profit-seeking and incur costs when they invest in the technology, and they must reach a bargaining agreement on how to share the revenue for the technology to reach the market.
no code implementations • 1 Aug 2023 • Alex John London, Hoda Heidari
The prevailing discourse around AI ethics lacks the language and formalism necessary to capture the diverse ethical concerns that emerge when AI systems interact with individuals.
no code implementations • 27 May 2023 • Michael Feffer, Hoda Heidari, Zachary C. Lipton
With Artificial Intelligence systems increasingly applied in consequential domains, researchers have begun to ask how these systems ought to act in ethically charged situations where even humans lack consensus.
no code implementations • 26 Mar 2023 • Anna Kawakami, Amanda Coston, Haiyi Zhu, Hoda Heidari, Kenneth Holstein
AI-based decision-making tools are rapidly spreading across a range of real-world, complex domains like healthcare, criminal justice, and child welfare.
no code implementations • 28 Jan 2023 • Hoda Heidari, Solon Barocas, Jon Kleinberg, Karen Levy
Prior work has provided strong evidence that, within organizational settings, teams that bring a diversity of information and perspectives to a task are more effective than teams that do not.
no code implementations • 30 Jun 2022 • Amanda Coston, Anna Kawakami, Haiyi Zhu, Ken Holstein, Hoda Heidari
Recent research increasingly brings to question the appropriateness of using predictive tools in complex, real-world tasks.
no code implementations • 13 May 2022 • Valerie Chen, Umang Bhatt, Hoda Heidari, Adrian Weller, Ameet Talwalkar
A practitioner may receive feedback from an expert at the observation- or domain-level, and convert this feedback into updates to the dataset, loss function, or parameter space.
no code implementations • 22 Apr 2022 • Charvi Rastogi, Liu Leqi, Kenneth Holstein, Hoda Heidari
To illustrate how our taxonomy can be used to investigate complementarity, we provide a mathematical aggregation framework to examine enabling conditions for complementarity.
no code implementations • 21 Apr 2022 • Nil-Jana Akpinar, Manish Nagireddy, Logan Stapleton, Hao-Fei Cheng, Haiyi Zhu, Steven Wu, Hoda Heidari
This stylized setup offers the distinct capability of testing fairness interventions beyond observational data and against an unbiased benchmark.
no code implementations • 12 Dec 2021 • Keegan Harris, Valerie Chen, Joon Sik Kim, Ameet Talwalkar, Hoda Heidari, Zhiwei Steven Wu
While the decision maker's problem of finding the optimal Bayesian incentive-compatible (BIC) signaling policy takes the form of optimization over infinitely-many variables, we show that this optimization can be cast as a linear program over finitely-many regions of the space of possible assessment rules.
1 code implementation • 12 Jul 2021 • Keegan Harris, Daniel Ngo, Logan Stapleton, Hoda Heidari, Zhiwei Steven Wu
In settings where Machine Learning (ML) algorithms automate or inform consequential decisions about people, individual decision subjects are often incentivized to strategically modify their observable attributes to receive more favorable predictions.
no code implementations • NeurIPS 2021 • Keegan Harris, Hoda Heidari, Zhiwei Steven Wu
In particular, we consider settings in which the agent's effort investment today can accumulate over time in the form of an internal state - impacting both his future rewards and that of the principal.
1 code implementation • 2 Jun 2021 • David Lindner, Hoda Heidari, Andreas Krause
To capture the long-term effects of ML-based allocation decisions, we study a setting in which the reward from each arm evolves every time the decision-maker pulls that arm.
no code implementations • 21 Jan 2021 • Hoda Heidari, Jon Kleinberg
We develop a dynamic model for allocating such opportunities in a society that exhibits bottlenecks in mobility; the problem of optimal allocation reflects a trade-off between the benefits conferred by the opportunities in the current generation and the potential to elevate the socioeconomic status of recipients, shaping the composition of future generations in ways that can benefit further from the opportunities.
Computers and Society Physics and Society
no code implementations • 8 Nov 2019 • Mohammad Yaghini, Andreas Krause, Hoda Heidari
Our family of fairness notions corresponds to a new interpretation of economic models of Equality of Opportunity (EOP), and it includes most existing notions of fairness as special cases.
1 code implementation • 4 Mar 2019 • Hoda Heidari, Vedant Nanda, Krishna P. Gummadi
Most existing notions of algorithmic fairness are one-shot: they ensure some form of allocative equality at the time of decision making, but do not account for the adverse impact of the algorithmic decisions today on the long-term welfare and prosperity of certain segments of the population.
no code implementations • 10 Sep 2018 • Hoda Heidari, Michele Loi, Krishna P. Gummadi, Andreas Krause
In this respect, our work serves as a unifying moral framework for understanding existing notions of algorithmic fairness.
no code implementations • 2 Jul 2018 • Till Speicher, Hoda Heidari, Nina Grgic-Hlaca, Krishna P. Gummadi, Adish Singla, Adrian Weller, Muhammad Bilal Zafar
Further, our work reveals overlooked tradeoffs between different fairness notions: using our proposed measures, the overall individual-level unfairness of an algorithm can be decomposed into a between-group and a within-group component.
no code implementations • NeurIPS 2018 • Hoda Heidari, Claudio Ferrari, Krishna P. Gummadi, Andreas Krause
We draw attention to an important, yet largely overlooked aspect of evaluating fairness for automated decision making systems---namely risk and welfare considerations.
1 code implementation • 7 Jun 2017 • Richard Berk, Hoda Heidari, Shahin Jabbari, Matthew Joseph, Michael Kearns, Jamie Morgenstern, Seth Neel, Aaron Roth
We introduce a flexible family of fairness regularizers for (linear and logistic) regression problems.
no code implementations • 27 Mar 2017 • Richard Berk, Hoda Heidari, Shahin Jabbari, Michael Kearns, Aaron Roth
Methods: We draw on the existing literatures in criminology, computer science and statistics to provide an integrated examination of fairness and accuracy in criminal justice risk assessments.