Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
HAI Weekly Seminar with Jeannette Bohg | Stanford HAI

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
    • Subscribe to Email
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
Your browser does not support the video tag.
eventSeminar

HAI Weekly Seminar with Jeannette Bohg

Status
Past
Date
Wednesday, April 06, 2022 10:00 AM - 11:00 AM PST/PDT
Location
Virtual
Share
Link copied to clipboard!
Event Contact
Kaci Peel
kpeel@stanford.edu

Related Events

Tom Mitchell | The History of Machine Learning
Feb 23, 202612:00 PM - 1:00 PM
February
23
2026

How did we get to today’s technology which now supports a trillion dollar AI industry? What were the key scientific breakthroughs? What were the surprises and dead-ends along the way...

Event

Tom Mitchell | The History of Machine Learning

Feb 23, 202612:00 PM - 1:00 PM

How did we get to today’s technology which now supports a trillion dollar AI industry? What were the key scientific breakthroughs? What were the surprises and dead-ends along the way...

Gaidi Faraj, Lofred Madzou | Nurturing Africa’s AI Leaders through Math Olympiad
SeminarFeb 25, 202612:00 PM - 1:15 PM
February
25
2026

The African Olympiad Academy is a world-class high school dedicated to training Africa’s most promising students in mathematics, science, and artificial intelligence through olympiad-based pedagogy.

Seminar

Gaidi Faraj, Lofred Madzou | Nurturing Africa’s AI Leaders through Math Olympiad

Feb 25, 202612:00 PM - 1:15 PM

The African Olympiad Academy is a world-class high school dedicated to training Africa’s most promising students in mathematics, science, and artificial intelligence through olympiad-based pedagogy.

Zoë Hitzig | How People Use ChatGPT
Mar 09, 202612:00 PM - 1:00 PM
March
09
2026

Despite the rapid adoption of LLM chatbots, little is known about how they are used. We approach this question theoretically and empirically, modeling a user who chooses whether to complete a task herself, ask the chatbot for information that reduces decision noise, or delegate execution to the chatbot...

Event

Zoë Hitzig | How People Use ChatGPT

Mar 09, 202612:00 PM - 1:00 PM

Despite the rapid adoption of LLM chatbots, little is known about how they are used. We approach this question theoretically and empirically, modeling a user who chooses whether to complete a task herself, ask the chatbot for information that reduces decision noise, or delegate execution to the chatbot...

On the Role of Vision, Touch and Sound for Robustness and Generalisability in Robotic Manipulation

Making Sense of Vision and Touch: Self-Supervised Learning of Multimodal Representations for Contact-Rich Tasks

Detect, Reject, Correct: Crossmodal Compensation of Corrupted Sensors

DiffImpact: Differentiable Rendering and Identification of Impact Sounds

Learning contact-rich, robotic manipulation skills is a challenging problem due to the high-dimensionality of the state and action space as well as uncertainty from noisy sensors and inaccurate motor control. In this research, Bohg and team explore what representations of raw perceptual data enable a robot to better learn and perform these skills. Specifically for manipulation robots, the sense of touch is essential yet it is non-trivial to manually design a robot controller that combines different sensing modalities that have very different characteristics.

Bohg will present the set of research work that explores the question of how to best fuse the information from vision and touch for contact-rich manipulation tasks. While deep reinforcement learning has shown success in learning control policies for high-dimensional inputs, these algorithms are generally intractable to deploy on real robots due to sample complexity. Bohg and team use self-supervision to learn a compact and multimodal representation of visual and haptic sensory inputs, which can then be used to improve the sample efficiency of policy learning. Bohg presents experiments on a peg insertion task where the learned policy generalizes over different geometry, configurations, and clearances, while being robust to external perturbations. The team also shows how exploiting multiple modalities helps to compensate for corrupted sensory data in one of the modalities.

Another modality that has been under-explored in robotic manipulation is sound. Rigid objects make distinctive sounds during manipulation. These sounds are a function of object features, such as shape and material, and of contact forces during manipulation. Being able to infer from sound an object's acoustic properties, how it is being manipulated, and what events it is participating in could augment and complement what robots can perceive from vision, especially in case of occlusion, low visual resolution, poor lighting, or blurred focus. Bohg will present a fully differentiable model for sounds rigid objects make during impacts, based on physical principles of impact forces, rigid object vibration, and other acoustic effects. Its differentiability enables gradient-based, efficient joint inference of acoustic properties of the objects and characteristics and timings of each individual impact. Bohg will conclude this talk with a discussion of appropriate representations for multimodal sensory data.

Jeannette BoghJeannette Bohg

Assistant Professor of Computer Science, Stanford University

Professor for Robotics and part of the Stanford AI lab within the Computer Science Department of Stanford University. Bohg is also directing the Interactive Perception and Robot Learning Lab, and enjoys research at the intersection of Robotics, Machine Learning and Computer Vision. No tweets available.