2024 Hoffman-Yee Grant Recipients
Six Stanford research teams have received funding to solve some of the most challenging problems in AI.
A human-centered foundation model for cells to accelerate drug discovery and personalized treatment
Imagine if we could build digital twins of your cells and simulate how they would respond to a drug treatment before exposing you to the drug while considering your sex, age, and comorbidities. The key to such a medical future requires developing end-to-end frameworks for cell modeling. Recent scientific and technological advances present a historic opportunity to make unprecedented progress in our fight against human disease. Specifically, new sources of biomedical data are rapidly becoming available, and new AI techniques make it possible to understand these massive datasets. We aim to (1) create a multimodal foundation model for cells capable of capturing function and state across human tissues and individuals, (2) develop an intuitive chat model interface to augment the ability of biologists to use and understand it, and (3) showcase its capacities by modeling cells affected by the menstrual cycle to answer critical questions in women’s health, with an initial focus on cardiovascular disease management.
NAME | ROLE | SCHOOL | DEPARTMENTS |
---|---|---|---|
Emma Lundberg | Lead PI | Engineering | Bioengineering |
Russ Altman | Co-PI | Engineering | Bioengineering |
Jure Leskovec | Co-PI | Engineering | Computer Science |
Stephen Quake | Co-PI | Engineering | Bioengineering |
Serena Yeung-Levy | Co-PI | Medicine | Biomedical Data Science |
Back to top of page |
A Multidimensional Odyssey into the Human Mind
The human brain, a complex organ influenced by psychological, biological, environmental, and physical factors, remains a significant challenge in neuroscience. Despite advances in understanding its structure, function, and biochemistry, a comprehensive, data-driven model integrating these diverse aspects has yet to be realized. Existing research has made strides with brain foundation models using MRI data and multi-modal approaches, but gaps remain in integrating various modalities and domains. This project aims to develop a “Brain World Model,” named “Brain-Bind,” by unifying datasets that reflect different aspects of brain health. This model will be trained in an end-to-end manner, harmonizing data sources, developing modality-specific encoders, and employing self-supervised learning to integrate non-imaging data with MRI-derived representations. The resulting model is intended to enhance diagnostic precision, inform personalized care, and contribute to a deeper understanding of neurological and cognitive processes.
NAME | ROLE | SCHOOL | DEPARTMENTS |
---|---|---|---|
Ehsan Adeli | Lead PI | Medicine | Psychiatry and Behavioral Sciences |
Akshay Chaudhari | Co-PI | Medicine | Radiology |
Anshul Kundaje | Co-PI | Medicine | Genetics |
Fei-Fei Li | Co-PI | Engineering | Computer Science |
Feng Vankee Lin | Co-PI | Medicine | Psychiatry and Behavioral Sciences |
Killian Pohl | Co-PI | Medicine | Psychiatry and Behavioral Sciences |
Jiajun Wu | Co-PI | Engineering | Computer Science |
Dan Yamins | Co-PI | Humanities and Sciences | Psychology |
Back to top of page |
Data in the Age of Generative AI
Massive datasets are the cornerstone for developing large language models (LLMs) and other generative AI. However, these datasets have also sparked debates regarding generative AI, highlighted by several copyright disputes involving OpenAI. This proposal is dedicated to exploring critical aspects of data creation and attribution for generative AI. Our approach is three-fold: Firstly, we aim to establish guiding principles and scaling laws for assembling datasets tailored for training and aligning LLMs and other generative AI, ensuring their responsible application in various sectors. Secondly, we plan to develop scalable methods to trace the generative AI outputs to specific training data, enabling data attribution and valuation. Finally, we will investigate how to effectively use and monitor synthetic data produced by generative AI. These goals are closely linked and mutually reinforcing, with successful data attribution and synthetic data methods informing and improving the strategies for dataset design. Throughout our project, we will ground our research with real-world legal and policy considerations and high-impact applications in law and medicine.
NAME | ROLE | SCHOOL | DEPARTMENTS |
---|---|---|---|
James Zou | Lead PI | Medicine | Biomedical Data Science |
Surya Ganguli | Co-PI | Humanities and Sciences | Applied Physics |
Tatsunori Hashimoto | Co-PI | Engineering | Computer Science |
Daniel Ho | Co-PI | Law | Law |
Curt Langlotz | Co-PI | Medicine | Radiology |
Percy Liang | Co-PI | Engineering | Computer Science |
Mark Lemley | Co-PI | Law | Law |
Megan Ma | Senior/Lead Research Scholar | Law | Law |
Julian Nyarko | Co-PI | Law | Law |
Christopher Ré | Co-PI | Engineering | Computer Science |
Ellen Vitercik | Co-PI | Engineering | Management Science and Engineering |
Back to top of page |
Evo: A foundational model for generative genomics
DNA encodes the fundamental language for all living organisms. Recently, large language models have been used to learn this mysterious biological language to unlock a better understanding of this blueprint of life. Yet, learning from DNA has its distinct challenges over natural language - it’s extremely long, with the human genome over 3 billion nucleotides in length. It’s also highly sensitive to small changes, where a single point mutation can mean the difference between having a disease or not. Overcoming these technical challenges of modeling long sequences in DNA can lead to a deeper understanding of human disease, the creation of novel therapeutics, and the possibility to engineer life itself.
This project aims to develop a new line of long sequence language models that can reproduce the organization of DNA sequences from the molecular to the whole genome scale. We will build on the Hyena architecture, an efficient long sequence model that leverages breakthroughs in deep signal processing and scales sub-quadratically with the length of data. We will extend the long convolutions of Hyena with a bidirectional and diffusion training paradigm. This approach will enable the modeling and design of DNA sequences from scratch as well as with an “infilling” ability, allowing both greater control and the mimicking of evolution’s process of continuous updating. Our team is dedicated to leading the ethical development of DNA sequence modeling and design, and to bring the innovation of AI systems for the betterment of human health.
NAME | ROLE | SCHOOL | DEPARTMENTS |
---|---|---|---|
Brian Hie | Lead PI | Engineering | Chemical Engineering |
Christopher Ré | Co-PI | Engineering | Computer Science |
Stefano Ermon | Co-PI | Engineering | Computer Science |
Stephen Baccus | Co-PI | Medicine | Neurobiology |
Euan Ashley | Co-PI | Medicine | Cardiovascular Medicine |
Back to top of page |
Integrating Intelligence: Building Shared Conceptual Grounding for Interacting with Generative AI
Visual media, in the form of images, video, animation, and 3D virtual environments are now central to the way people communicate stories, ideas and information. Yet, creating such visual content is challenging as it requires significant visual design expertise. The promise of modern generative AI tools is that they will assist users in creating production-quality visual content from a simple text prompt describing what the user wants. But current black-box AI are difficult to work with; the AI often misinterprets the intent of the user and users lack a predictive conceptual model for what the AI will produce given an input prompt. This mutual lack of a theory of mind leads to a collaboration by trial-and-error, where the user repeatedly tries different prompts hoping to find one that will produce the desired output.
In this project we take major steps towards allowing both entities (humans and AI) to develop a shared conceptual grounding that allows each to simulate how the other might operate given an input task. To this end, we focus on two key objectives. (1) First we will identify the concepts and mental processes expert human creators use when they make visual content. Such experts often convey their design intentions via natural language and sketches. We will analyze these linguistic and sketch outputs to identify common patterns in their workflows and we will investigate how such creators communicate with other human collaborators to establish common ground, repair misunderstandings, etc. (2) Second, we will build new generative AI tools that internally respect the ways humans mentally organize creation processes and workflows. For this objective, we will develop new algorithms and methods that incorporate the concepts humans use within generative AI models.
Ultimately we envision human creators collaborating with generative AI tools using a combination of natural language, example content and code snippets, in a turn-taking fashion to produce the desired content. Importantly, both the human and AI will communicate with one another through a shared understanding of the concepts and mental processes relevant to the creation task. We will validate our approach with human-subject experiments that examine how our generative AI tools lower usability barriers for creating production-quality visual content. With our generative AI tools we aim to democratize visual content creations so that users of all skill levels can easily express their ideas and tell their stories using visual media.
NAME | ROLE | SCHOOL | DEPARTMENTS |
---|---|---|---|
Maneesh Agrawala | Lead PI | Engineering | Computer Science |
Judith Fan | Co-PI | Humanities and Sciences | Psychology |
Kayvon Fatahalian | Co-PI | Engineering | Computer Science |
Tobi Gerstenberg | Co-PI | Humanities and Sciences | Psychology |
Nick Haber | Co-PI | Education | Graduate School of Education |
Hari Subramonyam | Co-PI | Education | Graduate School of Education |
Jiajun Wu | Co-PI | Engineering | Computer Science |
Back to top of page |
Leveraging Technology to Improve Police-Community Relations: The Promise of Large-Scale Analysis of Body-Worn Camera Footage
Police body-worn cameras have been at the center of police reform efforts over the past decade. Yet the vast majority of the footage generated by those cameras is never examined, undermining the camera’s utility as a tool for accountability and improving interactions between the police and community members. We are harnessing artificial intelligence (AI) and large language models (LLMs) to unlock the research potential of body-worn camera footage to better understand the nature of law enforcement’s encounters with the public. In turn, leveraging the resulting insights could fuel both the development and systematic evaluation of officer trainings and other institutional interventions designed to improve policing. To advance these goals, we are building AI tools and state-of-the-art infrastructure on Stanford’s campus to receive, secure, and process police footage for the purpose of conducting research aimed at improving relations between the police and the public. This includes body-worn camera footage of routine vehicle stops from several law enforcement agencies. We will begin by using this data to evaluate the effectiveness of a state-wide legal intervention designed to improve police-community interactions during vehicle stops, eradicate racial disparities in those stops, and increase trust. We will also explore the effectiveness of training and other methods to reduce escalation. In a world where police departments are increasingly utilizing AI to fight crime, we see the value of harnessing AI to make accessible a largely untapped source of data to improve police-community relations – a goal that can be lauded by the police and the policed alike. Using AI in this way could be crucial to advancing solutions to one of society’s most pressing problems: how can we reimagine public safety?
NAME | ROLE | SCHOOL | DEPARTMENTS |
---|---|---|---|
Jennifer Eberhardt | Lead PI | Graduate School of Business | Graduate School of Business |
Ralph Banks | Co-PI | Law | Law |
Dan Jurafsky | Co-PI | Humanities and Sciences | Linguistics |
Benoit Monin | Co-PI | Graduate School of Business | Graduate School of Business |
Back to top of page |