Responsible AI | The 2026 AI Index Report | Stanford HAI
Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
    • Subscribe to Email
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
Navigate
  • About
  • Events
  • AI Glossary
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

03

Responsible AI

Computer Vision

This chapter examines responsible AI across the dimensions of safety, fairness, transparency, and governance, and the measurement gaps that persist. 

See Chapter 4

All Chapters

  • Back to Overview
  • 01Research and Development
  • 02Technical Performance
  • 03Responsible AI
  • 04Economy
  • 05Science
  • 06Medicine
  • 07Education
  • 08Policy and Governance
  • 09Public Opinion

1. Responsible AI benchmarking is increasing, but is not keeping up with AI advances and deployments.

Almost all leading frontier model developers report results on capability benchmarks like MMLU and SWE-bench, but reporting on responsible AI benchmarks remains sparse. Documented AI incidents continued to rise, with the AI Incident Database recording 362 in 2025, up from 233 in 2024.

2. AI models struggle to tell the difference between knowledge and belief.

In a new accuracy benchmark, hallucination rates across 26 top models range from 22% to 94%.  GPT-4o's accuracy dropped from 98.2% to 64.4%, and DeepSeek R1 fell from over 90% to 14.4%. When a false statement is presented as something another person believes, models handle it well. When the same false statement is presented as something a user believes, performance collapses.

3. Organizations are formalizing responsible AI work, but knowledge and budget gaps still slow adoption.

AI-specific governance roles grew 17% in 2025, and the share of businesses with no responsible AI policies in place fell sharply from 24% to 11%. The main obstacles to implementation remain gaps in knowledge (59%), budget constraints (48%), and regulatory uncertainty (41%).

4. The mix of regulations shaping responsible AI practices is shifting toward AI-specific frameworks and technical standards.

GDPR remains the most cited regulatory influence but slipped from 65% in 2024 to 60% in 2025. New entries in 2025 include ISO/IEC 42001, an AI management system standard, cited by 36% of respondents, and the NIST AI Risk Management Framework at 33%. The share of organizations reporting no regulatory influence at all fell from 17% to 12%.

5. AI works best in English, and the gap is wider than global benchmarks suggest.

On HELM Arabic, a regionally developed model for the Arabic language, outscored GPT-5.1 and Gemini 2.5 Flash. The gap widens at the dialect level. On a Slovenian commonsense reasoning test, several leading models lost close to half their accuracy when tested in a regional dialect rather than the standard language.

6. AI companies grew less transparent this year.

After rising on the Foundation Model Transparency Index from 37 to 58 between 2023 and 2024,  the average score dropped to 40 in 2025. Major gaps persist in disclosure around training data, compute resources, and post-deployment impact.

7. AI models perform well on safety tests under normal conditions, but their defenses weaken under deliberate attack.

On the AILuminate benchmark, several frontier models received “Very Good” or “Good” safety ratings under standard use. When tested against jailbreak attempts using adversarial prompts, safety performance dropped across all models tested.

8. Responsible AI dimensions such as safety, fairness, and privacy are at odds with one another, and the tradeoffs are not well understood.

Recent empirical studies found that training techniques aimed at improving one responsible AI dimension consistently degraded others.


Support the Stanford Institute for Human-Centered AI (HAI) in our mission to advance ethical and impactful advancements in artificial intelligence.

Your support helps foster research, education, policy, and collaboration across diverse fields. Whether you are an individual, a corporation, a foundation, or a family office, together we can ensure AI serves humanity’s best interests.

Make a Gift to AI Index