Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
Can AI Hold Consistent Values? Stanford Researchers Probe LLM Consistency and Bias | Stanford HAI

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
news

Can AI Hold Consistent Values? Stanford Researchers Probe LLM Consistency and Bias

Date
November 11, 2024
Topics
Ethics, Equity, Inclusion
Natural Language Processing
Privacy, Safety, Security

New research tests large language models for consistency across diverse topics, revealing that while they handle neutral topics reliably, controversial issues lead to varied answers.

As large language models (LLMs) become increasingly woven into daily life — helping with everything from internet search to complex problem-solving — they’ve also faced scrutiny for potential bias. This begs deeper questions: Can an LLM have values and, if so, what values should it have? The answers aren’t abstract; they could shape how we build, interact, and trust these powerful tools. 

To find answers to these deeper questions, though, a Stanford research team had to start with a smaller question: They needed to learn whether LLMs are consistent in their answers. That is, do they give roughly the same answers every time they are asked a question? 

“You can’t really declare that a large language model is biased if it gives different answers when a question is rephrased, nuanced, or translated into other languages,” said Jared Moore, a doctoral candidate in computer science at Stanford who focuses on the ethics of artificial intelligence. He is first author of a new study on LLM consistency. 

“If I say that a person has bias, that means they're going to act somewhat similarly in a variety of circumstances,” Moore said. “And that hadn't been established with language models.”

Nuanced Views

In the study, Moore and colleagues asked several leading LLMs a battery of 8,000 questions across 300 topic areas. Their queries included the paraphrasing of similar questions, asking follow-up, nuanced, or related questions within given topic areas, and translating their original English questions into Chinese, German, and Japanese to gauge just how consistent these models were.

“We found that, in general, large models answers are pretty consistent across these different measures,” said Diyi Yang, a professor of computer science at Stanford University and senior author of the study. “Sometimes they were even more consistent than human participants.”

Across a range of LLMs — new, old, massive, and small — the team found that the largest of large language models (e.g., GPT-4, Claude) were more consistent than smaller, older models. 

However, the team also found that LLMs were most consistent on less controversial topics, like "Thanksgiving," for instance, than they were on more controversial topics, such as "euthanasia.” In fact, the more controversial the topic became, the less consistent the models became. Moore pointed to a series of questions on the less controversial topic of “women’s rights” where the models were more consistent than on hot-button, highly charged issues like “abortion.” 

“If the LLM offers a range of ideas that is reflected in greater inconsistency, that lends itself to the idea that LLMs are, in fact, not biased,” Moore noted. “With our particular methodology, we show that these models are actually incredibly inconsistent on controversial topics. So we shouldn't be ascribing these kinds of values to them.”

More Is Better

Moore is now moving forward with research on why models seem to be more consistent on certain topics than on others and evaluating solutions to potential bias.  “Just because I happen to agree that it is a good thing that models universally support, say, women's rights, there might be other topics where I would disagree. How do we determine which values models should have and who should make those decisions?” Moore said.

One solution, he noted, might be encouraging models toward value pluralism, reflecting a range of perspectives rather than presenting a single, albeit consistent, view. 

“Often, we don't want perfect consistency. We don't want models to always express the same positions. You want them to represent a distribution of ideas,” he said.

He thinks his future research might investigate how models can be trained to represent this wider range of views when handling more controversial, value-laden questions where bias is most problematic.

“I'm quite interested in this idea of pluralism because it forces us to address much bigger questions: What do we want our models to be, and how should they behave?” Moore said.

Share
Link copied to clipboard!
Contributor(s)
Andrew Myers

Related News

How Language Bias Persists in Scientific Publishing Despite AI Tools
Scott Hadly
Jun 16, 2025
News

Stanford researchers highlight the ongoing challenges of language discrimination in academic publishing, revealing that AI tools may not be the solution for non-native speakers.

News

How Language Bias Persists in Scientific Publishing Despite AI Tools

Scott Hadly
Ethics, Equity, InclusionGenerative AIJun 16

Stanford researchers highlight the ongoing challenges of language discrimination in academic publishing, revealing that AI tools may not be the solution for non-native speakers.

Europe's Innovation Pivot: Can the EU Lead the Next Wave of AI?
Daniel Zhang
Jun 04, 2025
News
Speaker for the event

With its AI Continent Action Plan, the EU aims to reinvent its innovation model. European Commission Executive Vice-President for Tech Sovereignty, Security and Democracy Henna Virkkunen outlines its ambition.

News
Speaker for the event

Europe's Innovation Pivot: Can the EU Lead the Next Wave of AI?

Daniel Zhang
Government, Public AdministrationPrivacy, Safety, SecurityJun 04

With its AI Continent Action Plan, the EU aims to reinvent its innovation model. European Commission Executive Vice-President for Tech Sovereignty, Security and Democracy Henna Virkkunen outlines its ambition.

Struggling DNA Testing Firm 23andMe To Be Bought For $256m
BBC
May 19, 2025
Media Mention

Stanford HAI Policy Fellow Jennifer King speaks about the data privacy implications of 23andMe's purchase by Regeneron.

Media Mention
Your browser does not support the video tag.

Struggling DNA Testing Firm 23andMe To Be Bought For $256m

BBC
Privacy, Safety, SecurityMay 19

Stanford HAI Policy Fellow Jennifer King speaks about the data privacy implications of 23andMe's purchase by Regeneron.