Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
Stanford Scholars Train Generative AI To Be Better Creative Collaborators | Stanford HAI
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
    • Subscribe to Email
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

news

Stanford Scholars Train Generative AI To Be Better Creative Collaborators

Date
March 10, 2026
Skilled comic artist creating comic book on computer

The team is building a shared “conceptual grounding” so that artists can steer models with precision.

The conversation around AI and art generally swings between two extremes: A flood of AI slop or the total automation of creative work. The more desirable approach may be an AI that behaves as a useful collaborator. 

But thus far, visual artists working with text-to-image tools confront frustrating basic hurdles in their abilities to direct AI. Ask an AI to create an image of a house? Not too difficult. Direct it to make the house red, with four front-facing windows, a chimney, and ivy covering the left side? Good luck.

Stanford computer science, cognitive psychology, and education scholars believe they can help AI better augment human creativity by teaching models and people to communicate ideas with each other. With funding from a Stanford Institute for Human-Centered AI (HAI) Hoffman-Yee Research Grant, the scholars are developing a shared conceptual grounding for humans to collaborate with generative AI on production-quality visual content ranging from illustrations to diagrams to animations.

“While the models seem amazing, they are terrible collaborators,” says Maneesh Agrawala, professor of computer science at Stanford and a co-principal investigator for the project. “Creators have no way of knowing what the AI will produce when given a certain text prompt. If you ask for a suburban single-family home, it generates a modern duplex.”

Authoring original content requires having opinions and constantly making choices, Agrawala explains. Humans and AI need a shared set of concepts so the nuance doesn’t get lost in translation. 

Deciphering the Human Creative Process

The Stanford team is approaching this problem from two directions. First, the scholars are running experiments to better understand how people collaborate to create visual content. They have conducted several studies of people performing creative tasks to analyze through chat logs and sketches how the participants communicate as they work together. 

“If we want to build AI systems that understand how humans think during creative projects, we should start by learning as much as we can from the way that people establish common conceptual ground with each other,” says Judith Fan, assistant professor of psychology at Stanford's School of Humanities and Sciences. “Not everyone talks or draws the same way, but they still expect to be understood.”  

Building AI Tools that Understand Creators

Second, the team is building open-source AI tools to apply the lessons learned about human creative communication. For example, ControlNet teaches text-to-image diffusion models about spatial composition, using two separate features, blocking and detailing, to mirror how artists begin with a rough sketch and then complete the detail of a drawing. Today’s models struggle to capture the idea of a pose or how objects should be arranged in a scene. With this tool, creators can guide models to a layout that matches their vision. 

Another tool called FramePack enables creators to generate 3D videos from a text prompt for multi-scene storytelling. This tool teaches models to prioritize scenes based on their importance to the overall story, similar to the way a human would work on the project.

A third innovation explores the power of neuro-symbolic AI, which combines neural networks with reasoning capabilities to increase transparency and overcome the limitations of “black box” AI. Using these principles, the team has developed a visual scene coding language that works from a natural language text prompt to produce lines of code, which are executed and rendered to create a 3D scene. Human creators can stay in the loop to inspect or edit the code and prompt the AI to update its program at any time.

Reimagining Education Content

The impact of a shared conceptual grounding between humans and AI promises to yield new applications in diverse fields, including design, simulation, animation, robotics, and education, says Agrawala. The research team is currently working with gaming platform Roblox to enable players to generate unique 3D objects from text prompts while imposing game restrictions (so, for example, players won’t be able to create weapons in a nonviolent game). 

More broadly, the scholars hope that one day human creators of all skill levels—from hobbyists and small business owners to visual experts—will have a friction-free way to express their ideas using a combination of natural language, example content, code snippets and other modalities. 

“We’re serious about equipping the broader creative community with the tools they need to communicate with AI effectively,” Fan says.

Want to learn more? Watch this research team discuss the latest findings during the recent Hoffman Yee Symposium at Stanford HAI. 

Share
Link copied to clipboard!
Contributor(s)
Nikki Goth Itoi

Related News

Who Decides How America Uses AI in War?
Curtis Langlotz, Amy Zegart, Michele Elam, Jennifer King, Russ Altman
Mar 30, 2026
News
image of drones connected by digital net

As artificial intelligence becomes central to national security, experts grapple with a technology that remains unpredictable, unregulated, and increasingly powerful.

News
image of drones connected by digital net

Who Decides How America Uses AI in War?

Curtis Langlotz, Amy Zegart, Michele Elam, Jennifer King, Russ Altman
Mar 30

As artificial intelligence becomes central to national security, experts grapple with a technology that remains unpredictable, unregulated, and increasingly powerful.

Stop Telling AI Your Secrets - 5 Reasons Why, And What To Do If You Already Overshared
ZD Net
Mar 25, 2026
Media Mention

"The ultimate problem is that you just can't control where the information goes, and it could leak out in ways that you just don't anticipate," says HAI Privacy and Data Policy Fellow Jennifer King.

Media Mention
Your browser does not support the video tag.

Stop Telling AI Your Secrets - 5 Reasons Why, And What To Do If You Already Overshared

ZD Net
Regulation, Policy, GovernanceGenerative AIMar 25

"The ultimate problem is that you just can't control where the information goes, and it could leak out in ways that you just don't anticipate," says HAI Privacy and Data Policy Fellow Jennifer King.

AI’s Growing Role as Scientific Peer Reviewer
Andrew Myers
Mar 25, 2026
News
illustration of a book turning into a computer with medical and science concepts

Stanford computer scientist James Zou is exploring how AI can accelerate scientific research and peer review. His finding: AI excels at spotting gaps, but judgment calls still need humans.

News
illustration of a book turning into a computer with medical and science concepts

AI’s Growing Role as Scientific Peer Reviewer

Andrew Myers
Sciences (Social, Health, Biological, Physical)Mar 25

Stanford computer scientist James Zou is exploring how AI can accelerate scientific research and peer review. His finding: AI excels at spotting gaps, but judgment calls still need humans.