Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
Cleaning Up Policy Sludge: An AI Statutory Research System | Stanford HAI

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
policyPolicy Brief

Cleaning Up Policy Sludge: An AI Statutory Research System

Date
June 18, 2025
Topics
Government, Public Administration
Read Paper
abstract

This brief introduces a novel AI tool that performs statutory surveys to help governments—such as the San Francisco City Attorney Office—identify policy sludge and accelerate legal reform.

Key Takeaways

  • Legal reform can get bogged down by policy sludge that is strewn about millions of words of statutes and regulations. Such sludge can make programs hard to administer for civil servants and even more difficult to navigate for the public.

  • Stanford RegLab developed the Statutory Research Assistant (STARA), a domain-informed AI system capable of performing accurate and comprehensive statutory surveys that help to identify and eliminate policy sludge.

  • As an illustration, RegLab partnered with the San Francisco City Attorney’s Office to identify all legislatively mandated reporting requirements, many of which are burdensome and can serve little purpose after decades. Based on the collaboration, the city attorney spearheaded a consultative process with city departments, culminating in a proposed ordinance to delete or consolidate over a third of these requirements.

  • AI systems like STARA enable researchers, advocates, attorneys, and government officials to gain a more comprehensive understanding of often opaque legal mandates, identify policy sludge, and accelerate meaningful reform efforts.

The Problem

There is a growing recognition that “policy sludge”—outdated, obsolete, or cumbersome legal requirements and regulations—can impede adaptable governance. But reforming a daunting volume of statutes, regulations, and codes can be challenging.

Consider five examples:

  • As a law professor in the 1970s, Ruth Bader Ginsburg hired an army of Columbia Law School students to cull through the United States Code for provisions that discriminated on the basis of sex. The final report, using 59 key words, was an “extensive, but not exhaustive” list which provided the blueprint for equal rights litigation.

  • In the 1980s, the U.S. Department of Justice under the Reagan administration tried to count federal crimes. After two years, they gave up. The responsible official said that one could “die and [be] resurrected three times,” and still not know the true number.

  • In 2021, California enacted a law that required all county recorder offices to identify and redact racist deed records. Such racial covenants, which prohibit people of particular races from residing on the property, are unenforceable and yet they persist. In Santa Clara County alone, that meant sifting through 80 million pages of deed documents dating back to the 1800s. Los Angeles recently spent $8 million for a contractor to use key words to find such covenants—a process expected to last over seven years.

  • Congressionally mandated reports are, according to political scientist Frank Fukuyama, a prime example of how “government is made inefficient by the layers of rules bureaucrats themselves are forced to labour under.” Congress has lost track of thousands of these reports, creating a congressional “black hole” that weighs down civil servants, with many reports producing little benefit. As Supreme Court Justice Neil Gorsuch noted, one report—on the Social Security Administration’s printing operations—took 95 employees over four months to complete. As the Congressional Research Service has conceded, there is no “search method that can obtain an exact accounting of all reports required,” given that the U.S. Code contains some 32 million words.

  • In 2024, San Francisco voters approved a ballot measure requiring the city to simplify its sprawling system of advisory bodies and commissions. The San Francisco Municipal Code, along with resolutions by the Board of Supervisors, totals 16 million words, and a civil grand jury found “there [was] no centralized list of commissions.”

  • The law runs across millions of statutes, regulations, deeds, and other documents. And sometimes, the problem facing policymakers, judges, and reformers is simply knowing what the law is.

In our paper, “What Is the Law? A System for Statutory Research (STARA) with Large Language Models,” we introduce an automated system that aims to address the unique challenges of statutory research by rapidly parsing and compiling legal provisions. It enables researchers, advocates, attorneys, and government officials to understand the full breadth of legislative mandates. Our work highlights the promise of using large language models (LLMs) to build domain-specific systems that perform well in complex fields such as statutory research, where they can help governments reduce policy sludge and pave the way for meaningful statutory reform.

The Solution

We developed the Statutory Research Assistant (STARA), a domain-informed AI system designed to automate statutory and regulatory research. STARA performs comprehensive statutory surveys, i.e., systematic compilations of legal provisions relevant to a given question or policy area. Unlike general-purpose tools, STARA exploits the capabilities of frontier AI models and a domain-specific architecture tailored to the structure of legal codes. It incorporates key elements such as hierarchical organization, cross-references, and definitions—unique features of legal codes that have made manual and automated approaches to statutory research challenging. Intuitively, STARA represents legal codes in the way the basics of statutory interpretation are taught to law students.

STARA’s research pipeline operates in three stages. First, it preprocesses and segments statutory codes into enumerable units, annotating each with relevant legal context drawn from headings, cross-references, definitions, and editorial notes. Then, it uses LLMs to reason about statutory language, classify provisions, and extract structured information. Finally, STARA can agentically organize, analyze, and report on the results of its statutory surveys.

This approach not only enables higher precision and recall than general-purpose tools, but it can also be adapted to parse diverse bodies of law—from the U.S. Code to state codes to city municipal codes. We benchmark STARA’s performance and show that it makes frontier LLMs as much as three times more accurate on complex statutory research tasks relative to a general-purpose AI system, making previously infeasible research efforts not just possible, but trivial.

Read Paper
Share
Link copied to clipboard!
Authors
  • Faiz Surani
    Faiz Surani
  • Lindsey A. Gailmard
    Lindsey A. Gailmard
  • Allison Casasola
    Allison Casasola
  • Varun Magesh
    Varun Magesh
  • Emily J. Robitschek
    Emily J. Robitschek
  • Christine Tsang
    Christine Tsang
  • Derek Ouyang
    Derek Ouyang
  • Dan Ho headshot
    Daniel E. Ho

Related Publications

Assessing the Implementation of Federal AI Leadership and Compliance Mandates
Jennifer Wang, Mirac Suzgun, Caroline Meinhardt, Daniel Zhang, Kazia Nowacki, Daniel E. Ho
Deep DiveJan 17, 2025
White Paper

This white paper assesses federal efforts to advance leadership on AI innovation and governance through recent executive actions and emphasizes the need for senior-level leadership to achieve a whole-of-government approach.

White Paper

Assessing the Implementation of Federal AI Leadership and Compliance Mandates

Jennifer Wang, Mirac Suzgun, Caroline Meinhardt, Daniel Zhang, Kazia Nowacki, Daniel E. Ho
Government, Public AdministrationRegulation, Policy, GovernanceDeep DiveJan 17

This white paper assesses federal efforts to advance leadership on AI innovation and governance through recent executive actions and emphasizes the need for senior-level leadership to achieve a whole-of-government approach.

Expanding Academia’s Role in Public Sector AI
Kevin Klyman, Caroline Meinhardt, Daniel Zhang, Elena Cryst, Russell Wald, Aaron Bao
Quick ReadDec 04, 2024
Issue Brief
Expanding Academias role in public sector ai

This brief analyzes the disparity between academia and industry in frontier AI research and presents policy recommendations for ensuring a stronger role for academia in public sector AI.

Issue Brief
Expanding Academias role in public sector ai

Expanding Academia’s Role in Public Sector AI

Kevin Klyman, Caroline Meinhardt, Daniel Zhang, Elena Cryst, Russell Wald, Aaron Bao
Government, Public AdministrationQuick ReadDec 04

This brief analyzes the disparity between academia and industry in frontier AI research and presents policy recommendations for ensuring a stronger role for academia in public sector AI.

Transparency of AI EO Implementation: An Assessment 90 Days In
Caroline Meinhardt, Kevin Klyman, Hamzah Daud, Christie M. Lawrence, Rohini Kosoglu, Daniel Zhang, Daniel E. Ho
Feb 22, 2024
Explainer
Your browser does not support the video tag.

The U.S. government has made swift progress and broadened transparency, but that momentum needs to be maintained for the next looming deadlines.

Explainer
Your browser does not support the video tag.

Transparency of AI EO Implementation: An Assessment 90 Days In

Caroline Meinhardt, Kevin Klyman, Hamzah Daud, Christie M. Lawrence, Rohini Kosoglu, Daniel Zhang, Daniel E. Ho
Regulation, Policy, GovernanceGovernment, Public AdministrationFeb 22

The U.S. government has made swift progress and broadened transparency, but that momentum needs to be maintained for the next looming deadlines.

Daniel E. Ho's Testimony Before the House Oversight and Accountability Cybersecurity, Information Technology, and Government Innovation Subcommittee
Daniel E. Ho
Dec 06, 2023
Testimony
Your browser does not support the video tag.

Daniel E. Ho's Testimony Before the House Oversight and Accountability Cybersecurity, Information Technology, and Government Innovation Subcommittee

Testimony
Your browser does not support the video tag.

Daniel E. Ho's Testimony Before the House Oversight and Accountability Cybersecurity, Information Technology, and Government Innovation Subcommittee

Daniel E. Ho
Privacy, Safety, SecurityGovernment, Public AdministrationDec 06

Daniel E. Ho's Testimony Before the House Oversight and Accountability Cybersecurity, Information Technology, and Government Innovation Subcommittee