Video available here.
Risk assessment instruments are used across the criminal justice system to estimate the probability of some future behavior given covariates. The estimated probabilities are then used in making decisions at the individual level. In the past, there has been controversy about whether the probabilities derived from group-level calculations can meaningfully be applied to individuals. Using Bayesian hierarchical models applied to a large longitudinal dataset from the court system in the state of Kentucky, we analyze variation in individual-level probabilities of failing to appear for court and the extent to which it is captured by covariates.
We find that individuals within the same risk group vary widely in their probability of the outcome. In practice, this means that allocating individuals to risk groups based on standard approaches to risk assessment, in large part, results in creating distinctions among individuals who are not meaningfully different in terms of their likelihood of the outcome. This is because uncertainty about the probability that any particular individual will fail to appear is large relative to the difference in average probabilities among any reasonable set of risk groups.
Kristian Lum is a senior staff machine learning researcher at Twitter in the Machine Learning Ethics, Transparency, and Accountability group. Prior to that she was a research professor at the University of Pennsylvania in the Department of Computer and Information Science and the lead statistician at the Human Rights Data Analysis Group. She was a founding member of the Executive Committee of the ACM Conference on Fairness, Accountability, and Transparency. Her research focuses on the responsible use of algorithmic decision-making, with an emphasis on evaluation of models for harmful impacts and mitigation techniques. In the past, her research has focused on the (un)fairness of predictive models used in the criminal justice system.
This talk will be recorded.