Abstract
Recent work on algorithmic fairness has largely focused on the fairness of discrete decisions, or classifications. While such decisions are often based on risk score models, the fairness of the risk models themselves has received considerably less attention. Risk models are of interest for a number of reasons, including the fact that they communicate uncertainty about the potential outcomes to users, thus representing a way to enable meaningful human oversight. Here, we address fairness desiderata for risk score models. We identify the provision of similar epistemic value to different groups as a key desideratum for risk score fairness, and we show how even fair risk scores can lead to unfair risk-based rankings. Further, we address how to assess the fairness of risk score models quantitatively, including a discussion of metric choices and meaningful statistical comparisons between groups. In this context, we also introduce a novel calibration error metric that is less sample size-biased than previously proposed metrics, enabling meaningful comparisons between groups of different sizes. We illustrate our methodology – which is widely applicable in many other settings – in two case studies, one in recidivism risk prediction, and one in risk of major depressive disorder (MDD) prediction.
Original language | English |
---|---|
Title of host publication | Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency |
Publisher | ACM |
Publication date | 2023 |
Pages | 817-829 |
ISBN (Electronic) | 979-8-4007-0192-4 |
DOIs | |
Publication status | Published - 2023 |
Event | 2023 ACM Conference on Fairness, Accountability, and Transparency - Chicago, United States Duration: 12 Jun 2023 → 15 Jun 2023 |
Conference
Conference | 2023 ACM Conference on Fairness, Accountability, and Transparency |
---|---|
Country/Territory | United States |
City | Chicago |
Period | 12/06/2023 → 15/06/2023 |
Keywords
- Algorithmic fairness
- Risk scores
- Ethics
- Ranking
- Recidivism
- Major depressive disorder
- Calibration
Fingerprint
Dive into the research topics of 'On (assessing) the fairness of risk score models'. Together they form a unique fingerprint.Press/Media
-
Eksperter bag nyt forskningsprojekt kalder på forsigtig udrulning af AI-modeller i det offentlige
Petersen, E. W., Ganz, M., Holm, S. H. & Feragen, A.
26/10/2023
1 item of Media coverage
Press/Media: Press / Media