Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
Can AI Hold Consistent Values? Stanford Researchers Probe LLM Consistency and Bias | Stanford HAI
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
    • Subscribe to Email
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

news

Can AI Hold Consistent Values? Stanford Researchers Probe LLM Consistency and Bias

Date
November 11, 2024
Topics
Ethics, Equity, Inclusion
Natural Language Processing
Privacy, Safety, Security

New research tests large language models for consistency across diverse topics, revealing that while they handle neutral topics reliably, controversial issues lead to varied answers.

As large language models (LLMs) become increasingly woven into daily life — helping with everything from internet search to complex problem-solving — they’ve also faced scrutiny for potential bias. This begs deeper questions: Can an LLM have values and, if so, what values should it have? The answers aren’t abstract; they could shape how we build, interact, and trust these powerful tools. 

To find answers to these deeper questions, though, a Stanford research team had to start with a smaller question: They needed to learn whether LLMs are consistent in their answers. That is, do they give roughly the same answers every time they are asked a question? 

“You can’t really declare that a large language model is biased if it gives different answers when a question is rephrased, nuanced, or translated into other languages,” said Jared Moore, a doctoral candidate in computer science at Stanford who focuses on the ethics of artificial intelligence. He is first author of a new study on LLM consistency. 

“If I say that a person has bias, that means they're going to act somewhat similarly in a variety of circumstances,” Moore said. “And that hadn't been established with language models.”

Nuanced Views

In the study, Moore and colleagues asked several leading LLMs a battery of 8,000 questions across 300 topic areas. Their queries included the paraphrasing of similar questions, asking follow-up, nuanced, or related questions within given topic areas, and translating their original English questions into Chinese, German, and Japanese to gauge just how consistent these models were.

“We found that, in general, large models answers are pretty consistent across these different measures,” said Diyi Yang, a professor of computer science at Stanford University and senior author of the study. “Sometimes they were even more consistent than human participants.”

Across a range of LLMs — new, old, massive, and small — the team found that the largest of large language models (e.g., GPT-4, Claude) were more consistent than smaller, older models. 

However, the team also found that LLMs were most consistent on less controversial topics, like "Thanksgiving," for instance, than they were on more controversial topics, such as "euthanasia.” In fact, the more controversial the topic became, the less consistent the models became. Moore pointed to a series of questions on the less controversial topic of “women’s rights” where the models were more consistent than on hot-button, highly charged issues like “abortion.” 

“If the LLM offers a range of ideas that is reflected in greater inconsistency, that lends itself to the idea that LLMs are, in fact, not biased,” Moore noted. “With our particular methodology, we show that these models are actually incredibly inconsistent on controversial topics. So we shouldn't be ascribing these kinds of values to them.”

More Is Better

Moore is now moving forward with research on why models seem to be more consistent on certain topics than on others and evaluating solutions to potential bias.  “Just because I happen to agree that it is a good thing that models universally support, say, women's rights, there might be other topics where I would disagree. How do we determine which values models should have and who should make those decisions?” Moore said.

One solution, he noted, might be encouraging models toward value pluralism, reflecting a range of perspectives rather than presenting a single, albeit consistent, view. 

“Often, we don't want perfect consistency. We don't want models to always express the same positions. You want them to represent a distribution of ideas,” he said.

He thinks his future research might investigate how models can be trained to represent this wider range of views when handling more controversial, value-laden questions where bias is most problematic.

“I'm quite interested in this idea of pluralism because it forces us to address much bigger questions: What do we want our models to be, and how should they behave?” Moore said.

Share
Link copied to clipboard!
Contributor(s)
Andrew Myers

Related News

AI Challenges Core Assumptions in Education
Shana Lynch
Feb 19, 2026
News

We need to rethink student assessment, AI literacy, and technology’s usefulness, according to experts at the recent AI+Education Summit.

News

AI Challenges Core Assumptions in Education

Shana Lynch
Education, SkillsGenerative AIPrivacy, Safety, SecurityFeb 19

We need to rethink student assessment, AI literacy, and technology’s usefulness, according to experts at the recent AI+Education Summit.

America's 250 Greatest Innovators: Celebrating The American Dream
Forbes
Feb 11, 2026
Media Mention

HAI Co-Director Fei-Fei Li named one of America's top 250 greatest innovators, alongside fellow Stanford affiliates Rodney Brooks, Carolyn Bertozzi, Daphne Koller, and Andrew Ng.

Media Mention
Your browser does not support the video tag.

America's 250 Greatest Innovators: Celebrating The American Dream

Forbes
Computer VisionGenerative AIFoundation ModelsEnergy, EnvironmentEthics, Equity, InclusionFeb 11

HAI Co-Director Fei-Fei Li named one of America's top 250 greatest innovators, alongside fellow Stanford affiliates Rodney Brooks, Carolyn Bertozzi, Daphne Koller, and Andrew Ng.

Smart Enough to Do Math, Dumb Enough to Fail: The Hunt for a Better AI Test
Andrew Myers
Feb 02, 2026
News
illustration of data and lines

A Stanford HAI workshop brought together experts to develop new evaluation methods that assess AI's hidden capabilities, not just its test-taking performance.

News
illustration of data and lines

Smart Enough to Do Math, Dumb Enough to Fail: The Hunt for a Better AI Test

Andrew Myers
Foundation ModelsGenerative AIPrivacy, Safety, SecurityFeb 02

A Stanford HAI workshop brought together experts to develop new evaluation methods that assess AI's hidden capabilities, not just its test-taking performance.