Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
A New Direction for Machine Learning in Criminal Law | Stanford HAI
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
    • Subscribe to Email
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

policyPolicy Brief

A New Direction for Machine Learning in Criminal Law

Date
December 01, 2021
Topics
Law Enforcement and Justice
Machine Learning
Read Paper
abstract

This brief proposes a machine learning approach to studying decision-making in the criminal legal system as a way to identify and reduce systemic inequalities.

Key Takeaways

  • Using machine learning to analyze decision-making in the criminal legal system could be a valuable way to identify discrimination and facilitate reconsideration of decisions where justice was inconsistently applied—but reconsideration is still a decision, and stakeholders in criminal law processes should consider whether and how machine learning should play a role in that decision.

  • We propose a two-pronged “Recon Approach” to use machine learning in criminal law: reconnaissance, where machine learning identifies patterns in a human decision-making process, and reconsideration, where machine learning then focuses on individual decision cases.

  • The Recon Approach is meant for scenarios where humans make discretionary decisions, there are records which describe decision factors, and those records are analyzable by a machine learning tool (e.g., a hearing transcript readable by a language-processing algorithm).

  • Policymakers should consider implementing stronger, clearer public record laws to ensure researchers have access to the necessary data to conduct these reviews.

Executive Summary

The American criminal legal system is rife with—and perpetuates— inequality. These discrimination problems across racial, socioeconomic, and other lines are well-documented, but studying the problem is still a resource-intensive process. Combing through court transcript hearings, manually categorizing case files, and other investigative techniques require hours of time, extensive manual labor, and research dollars that many civic and other organizations simply lack. Technology may be able to relieve some of this burden. 

Undoubtedly, many current applications of artificial intelligence (AI) technology have enabled, if not exacerbated, the discrimination problems in the criminal legal system. For instance, critics argue that because algorithms are trained on datasets reflecting centuries of racism (e.g., on arrest rates across racial groups), they tend to overestimate the risk of recidivism among defendants of color as compared to white defendants. There is justified concern about the use of technology in the criminal legal system and how it could focus on “technical fixes” for broad social and political problems in the criminal justice system and only make the problems worse.

In our new paper, we propose using machine learning (ML) to analyze decision-making in the criminal legal system. Instead of using ML to assess those being put through the system, we argue for using ML to analyze decisions that people in power have already made— thereby bringing increased transparency to those decisions, identifying patterns, and shining a light to help people see potential injustices. The aim is not to predict human behavior or replace human decision-making, but to better understand the factors that led to past decisions in the hopes of facilitating increased fairness and consistency in how criminal law is applied. We call it the “Recon Approach.”

Read Paper
Share
Link copied to clipboard!
Authors
  • Kristen Bell
    Kristen Bell
  • Jenny Hong
    Jenny Hong
  • Nick McKeown
    Nick McKeown
  • Catalin Voss
    Catalin Voss

Related Publications

Improving Transparency in AI Language Models: A Holistic Evaluation
Rishi Bommasani, Daniel Zhang, Tony Lee, Percy Liang
Quick ReadFeb 28, 2023
Issue Brief

This brief introduces Holistic Evaluation of Language Models (HELM) as a framework to evaluate commercial application of AI use cases.

Issue Brief

Improving Transparency in AI Language Models: A Holistic Evaluation

Rishi Bommasani, Daniel Zhang, Tony Lee, Percy Liang
Machine LearningFoundation ModelsQuick ReadFeb 28

This brief introduces Holistic Evaluation of Language Models (HELM) as a framework to evaluate commercial application of AI use cases.

Promoting Algorithmic Fairness in Clinical Risk Prediction
Stephen R. Pfohl, Agata Foryciarz, Nigam Shah
Quick ReadSep 09, 2022
Policy Brief

This brief examines the debate on algorithmic fairness in clinical predictive algorithms and recommends paths to safer, more equitable healthcare AI.

Policy Brief

Promoting Algorithmic Fairness in Clinical Risk Prediction

Stephen R. Pfohl, Agata Foryciarz, Nigam Shah
HealthcareMachine LearningEthics, Equity, InclusionQuick ReadSep 09

This brief examines the debate on algorithmic fairness in clinical predictive algorithms and recommends paths to safer, more equitable healthcare AI.

Brief Definitions of Key Terms in AI
Stanford HAI
Apr 01, 2022
Explainer

This explainer provides brief definitions for key terms associated with artificial intelligence, ranging from autonomous systems to deep learning and foundation models.

Explainer

Brief Definitions of Key Terms in AI

Stanford HAI
Machine LearningFoundation ModelsApr 01

This explainer provides brief definitions for key terms associated with artificial intelligence, ranging from autonomous systems to deep learning and foundation models.

Toward Responsible AI in Health Insurance Decision-Making
Michelle Mello, Artem Trotsyuk, Abdoul Jalil Djiberou Mahamadou, Danton Char
Quick ReadFeb 10, 2026
Policy Brief

This brief proposes governance mechanisms for the growing use of AI in health insurance utilization review.

Policy Brief

Toward Responsible AI in Health Insurance Decision-Making

Michelle Mello, Artem Trotsyuk, Abdoul Jalil Djiberou Mahamadou, Danton Char
HealthcareRegulation, Policy, GovernanceQuick ReadFeb 10

This brief proposes governance mechanisms for the growing use of AI in health insurance utilization review.