It's COMPASlicated: The Messy Relationship between RAI Datasets and Algorithmic Fairness BenchmarksDownload PDF

07 Jun 2021, 18:26 (modified: 07 Nov 2021, 09:25)NeurIPS 2021 Datasets and Benchmarks Track (Round 1)Readers: Everyone
Keywords: AI fairness, ai, machine learning, algorithmic fairness, risk assessment, benchmark, criminal justice, COMPAS
TL;DR: RAI datasets like COMPAS are ill-suited as benchmark datasets in algorithmic fairness due to biases/errors and limited relevance between RAI predictions on fairness in the real-world.
Abstract: Risk assessment instrument (RAI) datasets, particularly ProPublica’s COMPAS dataset, are commonly used in algorithmic fairness papers due to benchmarking practices of comparing algorithms on datasets used in prior work. In many cases, this data is used as a benchmark to demonstrate good performance without ac-counting for the complexities of criminal justice (CJ) processes. However, we show that pretrial RAI datasets can contain numerous measurement biases and errors, and due to disparities in discretion and deployment, algorithmic fairness applied to RAI datasets is limited in making claims about real-world outcomes.These reasons make the datasets a poor fit for benchmarking under assumptions of ground truth and real-world impact. Furthermore, conventional practices of simply replicating previous data experiments may implicitly inherit or edify normative positions without explicitly interrogating value-laden assumptions. Without con-text of how interdisciplinary fields have engaged in CJ research and context of how RAIs operate upstream and downstream, algorithmic fairness practices are misaligned for meaningful contribution in the context of CJ, and would benefit from transparent engagement with normative considerations and values related to fairness, justice, and equality. These factors prompt questions about whether benchmarks for intrinsically socio-technical systems like the CJ system can exist in a beneficial and ethical way.
Supplementary Material: zip
10 Replies