Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
Toward Political Neutrality in AI | Stanford HAI

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
policyPolicy Brief

Toward Political Neutrality in AI

Date
September 10, 2025
Topics
Democracy
Generative AI
Read Paper
abstract

This brief introduces a framework of eight techniques for approximating political neutrality in AI models.

Key Takeaways

  • True political neutrality in AI is impossible, but there are practical approximations of neutrality that developers can implement across different levels of AI systems.

  • We developed a framework for systematically evaluating political neutrality approximation techniques. The framework includes eight techniques across the output, system, and ecosystem level of AI systems.

  • More evaluations of political neutrality approximations are needed. Future research should focus on which approximations are currently used, which ones are feasible, and which ones are valued by users.

  • As a starting point for evaluation, we created a dataset to evaluate the output-level approximation techniques used by 10 current AI models. We find that, overall, open-source models exhibit more political bias and engage more readily with harmful content.

  • Policymakers and AI developers must shape AI systems that respect wide-ranging viewpoints while promoting fairness, user autonomy, and trust. To do so, they need to encourage transparency and interdisciplinary research on political neutrality approximations.

Executive Summary

Leading generative AI models have been reported to show political bias in individual instances – such as XAI’s Grok identifying as “MechaHitler” or Google’s Gemini depicting female popes — and in systematic ways.

Political bias is concerning because it is widespread and can influence users’ opinions and decisions. Recent research shows that AI-generated messages can influence people’s attitudes toward controversial issues such as gun control and climate action, and affect political decisions such as budget allocations. Politically biased AI systems may hinder people from independently forming opinions and making choices, a key pillar in liberal democracy. A commonly proposed solution to this challenge is making AI models politically neutral. However, true political neutrality in AI — meaning systems that are impartial and don’t favor particular political viewpoints — is theoretically and practically impossible.

In our paper “Political Neutrality in AI Is Impossible — But Here Is How to Approximate It,” we explain why this is the case and propose practical approximations of political neutrality that can reduce political bias and move us closer to achieving neutrality. We also test how today’s AI models respond to political content, and show how our framework can help evaluate and improve future language models.

Our work is a first step toward shifting the conversation on political bias away from impossible objectives and toward achievable approximations of political neutrality. These approximations allow AI developers to create systems that respect wide-ranging viewpoints while promoting fairness and user autonomy.

Why True Neutrality Is Impossible

Theoretically speaking, political neutrality is impossible. Neutrality is inherently subjective when what seems neutral to one person might seem biased to someone else. On the political spectrum, there is no neutral point, as moderate opinions that lie between left-leaning and right-leaning views are political positions in and of themselves. Evaluating political neutrality by assessing the intent or impact of an action is challenging as both are hard to measure.

Some argue that political neutrality is also currently technically impossible. In designing AI systems, humans make countless decisions about which data to use or how the system should respond — each of which can introduce biases. Even the information that AI models learn from, like the training data scraped from the internet or user inputs, often reflect existing biases. As a result, it is impossible to build an AI model without biased human input.

However, inspired by philosopher Joseph Raz, who observed that “neutrality [...] can be a matter of degree,” we argue that approximating political neutrality is not only possible but essential for promoting balanced AI interactions and mitigating user manipulation. Drawing on insights from related fields like sociology, political science, and philosophy, which have historically grappled with neutrality, bias, and representation, we developed a framework for approximating political neutrality in AI systems.

Read Paper
Share
Link copied to clipboard!
Authors
  • Jillian Fisher
    Jillian Fisher
  • Ruth E. Appel
    Ruth E. Appel
  • Yulia Tsvetkov
    Yulia Tsvetkov
  • Margaret E. Roberts
    Margaret E. Roberts
  • Jennifer Pan
    Jennifer Pan
  • Dawn Song
    Dawn Song
  • yejin choi
    Yejin Choi
Related
  • Whose Opinions Do Language Models Reflect?
    Shibani Santurkar, Esin Durmus, Faisal Ladhak, Cinoo Lee, Percy Liang, Tatsunori Hashimoto
    Quick ReadSep 20
    policy brief

    This brief introduces a quantitative framework that allows policymakers to evaluate the behavior of language models to assess what kinds of opinions they reflect.

Related Publications

Labeling AI-Generated Content May Not Change Its Persuasiveness
Isabel Gallegos, Dr. Chen Shani, Weiyan Shi, Federico Bianchi, Izzy Benjamin Gainsburg, Dan Jurafsky, Robb Willer
Quick ReadJul 30, 2025
Policy Brief

This brief evaluates the impact of authorship labels on the persuasiveness of AI-written policy messages.

Policy Brief

Labeling AI-Generated Content May Not Change Its Persuasiveness

Isabel Gallegos, Dr. Chen Shani, Weiyan Shi, Federico Bianchi, Izzy Benjamin Gainsburg, Dan Jurafsky, Robb Willer
Generative AIRegulation, Policy, GovernanceQuick ReadJul 30

This brief evaluates the impact of authorship labels on the persuasiveness of AI-written policy messages.

Simulating Human Behavior with AI Agents
Joon Sung Park, Carolyn Q. Zou, Aaron Shaw, Benjamin Mako Hill, Carrie J. Cai, Meredith Ringel Morris, Robb Willer, Percy Liang, Michael S. Bernstein
Quick ReadMay 20, 2025
Policy Brief

This brief introduces a generative AI agent architecture that can simulate the attitudes of more than 1,000 real people in response to major social science survey questions.

Policy Brief

Simulating Human Behavior with AI Agents

Joon Sung Park, Carolyn Q. Zou, Aaron Shaw, Benjamin Mako Hill, Carrie J. Cai, Meredith Ringel Morris, Robb Willer, Percy Liang, Michael S. Bernstein
Generative AIQuick ReadMay 20

This brief introduces a generative AI agent architecture that can simulate the attitudes of more than 1,000 real people in response to major social science survey questions.

How Persuasive is AI-Generated Propaganda?
Josh A. Goldstein, Jason Chao, Shelby Grossman, Alex Stamos, Michael Tomz
Quick ReadSep 03, 2024
Policy Brief

This brief presents the findings of an experiment that measures how persuasive AI-generated propaganda is compared to foreign propaganda articles written by humans.

Policy Brief

How Persuasive is AI-Generated Propaganda?

Josh A. Goldstein, Jason Chao, Shelby Grossman, Alex Stamos, Michael Tomz
DemocracyFoundation ModelsQuick ReadSep 03

This brief presents the findings of an experiment that measures how persuasive AI-generated propaganda is compared to foreign propaganda articles written by humans.

Demographic Stereotypes in Text-to-Image Generation
Federico Bianchi, Pratyusha Kalluri, Esin Durmus, Faisal Ladhak, Myra Cheng, Debora Nozza, Tatsunori Hashimoto, Dan Jurafsky, James Zou, Aylin Caliskan
Quick ReadNov 30, 2023
Policy Brief

This brief tests a variety of ordinary text prompts to examine how major text-to-image AI models encode a wide range of dangerous biases about demographic groups.

Policy Brief

Demographic Stereotypes in Text-to-Image Generation

Federico Bianchi, Pratyusha Kalluri, Esin Durmus, Faisal Ladhak, Myra Cheng, Debora Nozza, Tatsunori Hashimoto, Dan Jurafsky, James Zou, Aylin Caliskan
Generative AIFoundation ModelsEthics, Equity, InclusionQuick ReadNov 30

This brief tests a variety of ordinary text prompts to examine how major text-to-image AI models encode a wide range of dangerous biases about demographic groups.