Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
Natural Language Processing | Stanford HAI

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
Back to Natural Language Processing

All Work Published on Natural Language Processing

Understanding Social Reasoning in Language Models with Language Models
Kanishk Gandhi, Jan-Philipp Fränken, Tobias Gerstenberg, Noah Goodman
Sep 25, 2023
Research
Your browser does not support the video tag.

As Large Language Models (LLMs) become increasingly integrated into our everyday lives, understanding their ability to comprehend human mental states becomes critical for ensuring effective interactions. However, despite the recent attempts to assess the Theory-of-Mind (ToM) reasoning capabilities of LLMs, the degree to which these models can align with human ToM remains a nuanced topic of exploration. This is primarily due to two distinct challenges: (1) the presence of inconsistent results from previous evaluations, and (2) concerns surrounding the validity of existing evaluation methodologies. To address these challenges, we present a novel framework for procedurally generating evaluations with LLMs by populating causal templates. Using our framework, we create a new social reasoning benchmark (BigToM) for LLMs which consists of 25 controls and 5,000 model-written evaluations. We find that human participants rate the quality of our benchmark higher than previous crowd-sourced evaluations and comparable to expert-written evaluations. Using BigToM, we evaluate the social reasoning capabilities of a variety of LLMs and compare model performances with human performance. Our results suggest that GPT4 has ToM capabilities that mirror human inference patterns, though less reliable, while other LLMs struggle.

Understanding Social Reasoning in Language Models with Language Models

Kanishk Gandhi, Jan-Philipp Fränken, Tobias Gerstenberg, Noah Goodman
Sep 25, 2023

As Large Language Models (LLMs) become increasingly integrated into our everyday lives, understanding their ability to comprehend human mental states becomes critical for ensuring effective interactions. However, despite the recent attempts to assess the Theory-of-Mind (ToM) reasoning capabilities of LLMs, the degree to which these models can align with human ToM remains a nuanced topic of exploration. This is primarily due to two distinct challenges: (1) the presence of inconsistent results from previous evaluations, and (2) concerns surrounding the validity of existing evaluation methodologies. To address these challenges, we present a novel framework for procedurally generating evaluations with LLMs by populating causal templates. Using our framework, we create a new social reasoning benchmark (BigToM) for LLMs which consists of 25 controls and 5,000 model-written evaluations. We find that human participants rate the quality of our benchmark higher than previous crowd-sourced evaluations and comparable to expert-written evaluations. Using BigToM, we evaluate the social reasoning capabilities of a variety of LLMs and compare model performances with human performance. Our results suggest that GPT4 has ToM capabilities that mirror human inference patterns, though less reliable, while other LLMs struggle.

Natural Language Processing
Foundation Models
Sciences (Social, Health, Biological, Physical)
Your browser does not support the video tag.
Research
Stanford HAI Announces Hoffman-Yee Grants Recipients for 2024
Nikki Goth Itoi
Aug 21, 2024
Announcement

Six interdisciplinary research teams received a total of $3 million to pursue groundbreaking ideas in the field of AI.

Stanford HAI Announces Hoffman-Yee Grants Recipients for 2024

Nikki Goth Itoi
Aug 21, 2024

Six interdisciplinary research teams received a total of $3 million to pursue groundbreaking ideas in the field of AI.

Design, Human-Computer Interaction
Healthcare
Natural Language Processing
Machine Learning
Announcement
Meta’s New Llama 3.1 AI Model Is Free, Powerful, And Risky
WIRED
Jul 23, 2024
Media Mention

With the release of Meta's Llama 3.1, Director of CRFM and Senior Fellow at Stanford HAI Percy Liang comments on the potential audience shifts that could occur from other commercial AI tools to Llama 3.1.

Meta’s New Llama 3.1 AI Model Is Free, Powerful, And Risky

WIRED
Jul 23, 2024

With the release of Meta's Llama 3.1, Director of CRFM and Senior Fellow at Stanford HAI Percy Liang comments on the potential audience shifts that could occur from other commercial AI tools to Llama 3.1.

Generative AI
Natural Language Processing
Machine Learning
Foundation Models
Media Mention
Law Clerk vs. AI? Courthouse Test Highlights Judicial Curiosity
Bloomberg Law
Jul 03, 2024
Media Mention

Stanford HAI Senior Fellow Daniel E. Ho comments on his research on legal hallucinations in large language models and the viability of using similar models for judicial interpretation.

Law Clerk vs. AI? Courthouse Test Highlights Judicial Curiosity

Bloomberg Law
Jul 03, 2024

Stanford HAI Senior Fellow Daniel E. Ho comments on his research on legal hallucinations in large language models and the viability of using similar models for judicial interpretation.

Natural Language Processing
Foundation Models
Law Enforcement and Justice
Media Mention
AI can Outperform Humans in Writing Medical Summaries
Andrew Myers
Jun 03, 2024
News

A new study adapts large language models to summarize clinical documents, showing a promising path for AI to improve clinical workflows and patient care.

AI can Outperform Humans in Writing Medical Summaries

Andrew Myers
Jun 03, 2024

A new study adapts large language models to summarize clinical documents, showing a promising path for AI to improve clinical workflows and patient care.

Healthcare
Natural Language Processing
News
How Bias Hides in ‘Kitchen Sink’ Approaches to Data
Julian Nyarko
Andrew Myers
May 30, 2024
News

In risk modeling, AI researchers take a more-is-better approach to training data, but a new study argues that a less-is-more approach may be preferable.

How Bias Hides in ‘Kitchen Sink’ Approaches to Data

Julian Nyarko
Andrew Myers
May 30, 2024

In risk modeling, AI researchers take a more-is-better approach to training data, but a new study argues that a less-is-more approach may be preferable.

Natural Language Processing
Machine Learning
News
3
4
5
6
7