Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
Whose Opinions Do Language Models Reflect? | Stanford HAI

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
    • Subscribe to Email
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
policyPolicy Brief

Whose Opinions Do Language Models Reflect?

Date
September 20, 2023
Topics
Generative AI
Ethics, Equity, Inclusion
Read Paper
abstract

This brief introduces a quantitative framework that allows policymakers to evaluate the behavior of language models to assess what kinds of opinions they reflect.

Key Takeaways

  • Language models are shaped by a variety of inputs and opinions, from the individuals whose views are included in the training data to crowd workers who manually filter that data.

  • We found that language models fine-tuned with human feedback—meaning models that went through additional training with human input—were less representative of the general public’s opinions than models that were not fine-tuned.

  • It is possible to steer a language model toward the opinions of a particular demographic group by asking the model to respond as if it were a member of that group, but this can lead to undesirable side effects, such as exacerbating polarization and creating echo chambers.

  • We highlight the need for further research on the evaluation of language models that can help policymakers and regulators quantitatively assess language model behavior and compare it to human preferences and opinions.

Executive Summary

Since the November 2022 debut of ChatGPT, an Al chatbot developed by OpenAl, language models have been all over the news. But as people use chatbots-to write stories and look up recipes, to make travel plans and even further a real estate business-journalists, policymakers, and members of the public are increasingly paying attention to the important question of whose opinions these language models reflect. In particular, one emerging concern is that Al-generated text may be able to influence our views, including political beliefs, without our realizing it.

Language models, chatbots included, are shaped by a variety of data inputs. These inputs are provided by internet users in the form of training data (such as the authors of internet comments or blogs), crowd workers providing feedback on how to improve data or models (as OpenAl used in Kenya), and the developers themselves (who make high-level decisions regarding data collection and training). The data used to inform language models therefore represents a range of individuals and draws on a wide variety of opinions about sports, politics, culture, food, and many other topics. Meanwhile, language models are being asked subjective questions that have no clear right or wrong answer.

In our paper, "Whose Opinions Do Language Models Reflect?," we introduce a quantitative framework to answer this very question. The framework includes our development of a dataset to evaluate language models' alignment with 60 demographic groups in the United States, covering a diversity of topics. Using this framework, we find a major gap between the responses provided by language models and the views of demographic groups in the United States. We also discover a number of U.S. groups whose views are poorly reflected by current language models, such as people 65 years of age and older, widowed individuals, and people who regularly attend religious services. Our framework allows policymakers to quantitatively evaluate language models and serves as a reminder that issues of representation in language models should remain front of mind.

Read Paper
Share
Link copied to clipboard!
Authors
  • Shibani Santurkar
    Shibani Santurkar
  • Esin Durmus
    Esin Durmus
  • Faisal Ladhak
    Faisal Ladhak
  • Cinoo Lee
    Cinoo Lee
  • Percy Liang
    Percy Liang
  • Tatsunori Hashimoto
    Tatsunori Hashimoto

Related Publications

Moving Beyond the Term "Global South" in AI Ethics and Policy
Evani Radiya-Dixit, Angèle Christin
Quick ReadNov 19, 2025
Issue Brief

This brief examines the limitations of the term "Global South" in AI ethics and policy, and highlights the importance of grounding such work in specific regions and power structures.

Issue Brief

Moving Beyond the Term "Global South" in AI Ethics and Policy

Evani Radiya-Dixit, Angèle Christin
Ethics, Equity, InclusionInternational Affairs, International Security, International DevelopmentQuick ReadNov 19

This brief examines the limitations of the term "Global South" in AI ethics and policy, and highlights the importance of grounding such work in specific regions and power structures.

Toward Political Neutrality in AI
Jillian Fisher, Ruth E. Appel, Yulia Tsvetkov, Margaret E. Roberts, Jennifer Pan, Dawn Song, Yejin Choi
Quick ReadSep 10, 2025
Policy Brief

This brief introduces a framework of eight techniques for approximating political neutrality in AI models.

Policy Brief

Toward Political Neutrality in AI

Jillian Fisher, Ruth E. Appel, Yulia Tsvetkov, Margaret E. Roberts, Jennifer Pan, Dawn Song, Yejin Choi
DemocracyGenerative AIQuick ReadSep 10

This brief introduces a framework of eight techniques for approximating political neutrality in AI models.

Increasing Fairness in Medicare Payment Algorithms
Marissa Reitsma, Thomas G. McGuire, Sherri Rose
Quick ReadSep 01, 2025
Policy Brief

This brief introduces two algorithms that can promote fairer Medicare Advantage spending for minority populations.

Policy Brief

Increasing Fairness in Medicare Payment Algorithms

Marissa Reitsma, Thomas G. McGuire, Sherri Rose
Ethics, Equity, InclusionHealthcareQuick ReadSep 01

This brief introduces two algorithms that can promote fairer Medicare Advantage spending for minority populations.

Labeling AI-Generated Content May Not Change Its Persuasiveness
Isabel Gallegos, Dr. Chen Shani, Weiyan Shi, Federico Bianchi, Izzy Benjamin Gainsburg, Dan Jurafsky, Robb Willer
Quick ReadJul 30, 2025
Policy Brief

This brief evaluates the impact of authorship labels on the persuasiveness of AI-written policy messages.

Policy Brief

Labeling AI-Generated Content May Not Change Its Persuasiveness

Isabel Gallegos, Dr. Chen Shani, Weiyan Shi, Federico Bianchi, Izzy Benjamin Gainsburg, Dan Jurafsky, Robb Willer
Generative AIRegulation, Policy, GovernanceQuick ReadJul 30

This brief evaluates the impact of authorship labels on the persuasiveness of AI-written policy messages.