![A “COMPAS” That's Pointing in the Wrong Direction – Data Science W231 | Behind the Data: Humans and Values A “COMPAS” That's Pointing in the Wrong Direction – Data Science W231 | Behind the Data: Humans and Values](https://blogs.ischool.berkeley.edu/w231/files/2021/07/Screen-Shot-2021-07-16-at-4.22.36-PM-300x300.png)
A “COMPAS” That's Pointing in the Wrong Direction – Data Science W231 | Behind the Data: Humans and Values
![Rachel Thomas on Twitter: "@harini824 The Compas Recidivism Algorithm: - it's no more accurate than random people (Amazon Mechanical Turk) - it's a black box with 137 inputs but no more accurate Rachel Thomas on Twitter: "@harini824 The Compas Recidivism Algorithm: - it's no more accurate than random people (Amazon Mechanical Turk) - it's a black box with 137 inputs but no more accurate](https://pbs.twimg.com/media/EId_gOcU8AAdKfq.png)
Rachel Thomas on Twitter: "@harini824 The Compas Recidivism Algorithm: - it's no more accurate than random people (Amazon Mechanical Turk) - it's a black box with 137 inputs but no more accurate
Racial Bias and Gender Bias Examples in AI systems - Adolfo Eliazàt - Artificial Intelligence - AI News
Rachel Thomas on Twitter: "The Compas recidivism algorithm used in US courts has double the false positive rate (people rated high risk who do not reoffend) for Black defendants compared to white
![Data and Discretion: Why We Should Exercise Caution Around Using the COMPAS Algorithm in Court | Stanford Rewired Data and Discretion: Why We Should Exercise Caution Around Using the COMPAS Algorithm in Court | Stanford Rewired](https://cdn.stanfordrewired.com/wp-content/uploads/2020/09/Frame-9.jpg)