AI-powered EDGE Dance Animator Applies Generative AI to Choreography | Stanford HAI
Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
    • Subscribe to Email
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
Navigate
  • About
  • Events
  • AI Glossary
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

news

AI-powered EDGE Dance Animator Applies Generative AI to Choreography

Date
April 20, 2023
Topics
Design, Human-Computer Interaction
Machine Learning

AI analyzes the music’s rhythmic and emotional content and creates realistic dances that are also physically plausible — a real dancer could perform them.

Stanford University researchers have developed a generative AI model that can choreograph human dance animation to match any piece of music. It’s called Editable Dance GEneration (EDGE).

“EDGE shows that AI-enabled characters can bring a level of musicality and artistry to dance animation that was not possible before,” says Karen Liu, a professor of computer science who led a team that included two student collaborators, Jonathan Tseng and Rodrigo Castellon, in her lab.

The researchers believe that the tool will help choreographers design sequences and communicate their ideas to live dancers by visualizing 3D dance sequences. Key to the program’s advanced capabilities is editability. Liu imagines that EDGE could be used to create computer-animated dance sequences by allowing animators to intuitively edit any parts of dance motion.

For example, the animator can design specific leg movements of the character, and EDGE will “auto-complete” the entire body from that positioning in a way that is realistic, seamless, and physically plausible as well — a human could complete the moves. Above all, the moves are consistent with the animator’s choice of music.

Like other generative models for images and text — ChatGPT and DALL-E, for instance — EDGE represents a new tool for choreographic idea generation and movement planning. The editability means that dance artists and choreographers can iteratively refine their sequences move by move, position by position, adding specific poses at precise moments. EDGE then incorporates the additional details into the sequence automatically. In the near future, EDGE will allow users to input their own music and even demonstrate the moves themselves in front of a camera.

“We think it’s a really a fun and engaging way for everyone, not just dancers, to express themselves through movement and tap into their own creativity,” Liu says.

“With its ability to generate captivating dances in response to any music, we think EDGE represents a major milestone in the intersection of technology and movement,” adds Tseng. “It will unlock new possibilities for creative expression and physical engagement,” says Castellon.

The team has published a paper and will formally introduce EDGE at the Computer Vision and Pattern Recognition conference in Vancouver, British Columbia, in June.

Stanford HAI’s mission is to advance AI research, education, policy and practice to improve the human condition. Learn more. 

Share
Link copied to clipboard!
Contributor(s)
Andrew Myers

Related News

Stanford Study: AI Experts Are Optimistic About AI. The Rest of Us… Not So Much
KQED
Apr 13, 2026
Media Mention

Sha Sajadieh, AI Index Lead, comments on HAI's 2026 AI Index findings.

Media Mention
Your browser does not support the video tag.

Stanford Study: AI Experts Are Optimistic About AI. The Rest of Us… Not So Much

KQED
Workforce, LaborSciences (Social, Health, Biological, Physical)Design, Human-Computer InteractionEthics, Equity, InclusionApr 13

Sha Sajadieh, AI Index Lead, comments on HAI's 2026 AI Index findings.

Want To Understand The Current State Of AI? Check Out These Charts.
MIT Technology Review
Apr 13, 2026
Media Mention

"If you’re following AI news, you’re probably getting whiplash. AI is a gold rush. AI is a bubble. AI is taking your job. AI can’t even read a clock. The 2026 AI Index from Stanford University’s Institute for Human-Centered Artificial Intelligence, AI’s annual report card, comes out today and cuts through some of that noise."

Media Mention
Your browser does not support the video tag.

Want To Understand The Current State Of AI? Check Out These Charts.

MIT Technology Review
International Affairs, International Security, International DevelopmentEducation, SkillsRegulation, Policy, GovernanceMachine LearningWorkforce, LaborApr 13

"If you’re following AI news, you’re probably getting whiplash. AI is a gold rush. AI is a bubble. AI is taking your job. AI can’t even read a clock. The 2026 AI Index from Stanford University’s Institute for Human-Centered Artificial Intelligence, AI’s annual report card, comes out today and cuts through some of that noise."

How a HAI Seed Grant Helped Launch a Disease-Fighting AI Platform
Dylan Walsh
Mar 03, 2026
News

Stanford scientists in Senegal hunting for schistosomiasis—a parasitic disease infecting 200+ million people worldwide—used AI to transform local field work into satellite-powered disease mapping.

News

How a HAI Seed Grant Helped Launch a Disease-Fighting AI Platform

Dylan Walsh
Computer VisionHealthcareSciences (Social, Health, Biological, Physical)Machine LearningMar 03

Stanford scientists in Senegal hunting for schistosomiasis—a parasitic disease infecting 200+ million people worldwide—used AI to transform local field work into satellite-powered disease mapping.