Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
Computer Scientists Can’t Treat Social and Ethical Impacts as an Afterthought | Stanford HAI

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
    • Subscribe to Email
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
news

Computer Scientists Can’t Treat Social and Ethical Impacts as an Afterthought

Date
July 21, 2022
Topics
Education, Skills

A new National Academies of Science report argues that researchers must start projects with ethical review, working with stakeholders and experts from other fields.

When people sound alarms about ethical and social pitfalls in computing, especially artificial intelligence, they are often reacting to systems that are already in use. How should a social media platform handle algorithms that amplify hate speech and misinformation? Do systems that evaluate creditworthiness or job applications have hidden racial or gender biases? Does facial recognition jeopardize privacy?

But a new report from an advisory committee to the National Academies of Science, whose members include John Hennessy, the former president of Stanford and an advisor to Stanford HAI, argues that computer researchers and the institutions that fund them need to anticipate social and ethical risks long before they have a product.

If they don’t, the report warns, it may be too late.

“It is much easier to design a technology correctly from the start than it is to fix it later,” the report warns. “Failure to consider the consequences early in research increases the risk of adverse societal or ethical impacts.”

Read the full report, Fostering Responsible Computing Research: Foundations and Practices Report Release.

 

That may sound obvious, but the authors — including luminaries in computer science, social science, and philosophy — say it requires a broad rethink by the institutions that fund and carry out research: universities, corporations, professional societies, and the government.

In part, that means reaching out early to stakeholders as well as to experts in social sciences, ethics, and moral reasoning. It also means thinking early and hard about the unexpected ways that a new technology might be used or misused.

“One of the difficulties is that computer technologies, especially these foundation models, are universal technologies that can be used for all kinds of things that the developers never intended,” says Hennessy, a computer scientist and professor at Stanford who is also currently chairman of Alphabet, the parent company of Google. “You can’t prevent all the misuses, but you can at least provide a caveat that people can use as guidelines.”

Chain of Responsibility

The report cites the example of third-party “cookies,” the little markers that track actions of a user on a website. The original intent was simply to make transactions like online shopping easier, but cookies quickly became tools that data harvesters used to track users and their web activity. If researchers had considered privacy issues at the outset, Hennessey says, they could have built in more protection before cookies became a universal standard.

The open-ended potential of AI and other computer technology makes them different from innovations in most other fields. A new vaccine may have unwanted side effects, Hennessy notes, but it will only be used for a very targeted purpose. A new algorithm or piece of code becomes a tool that can be used for entirely new purposes.

“The chain of responsibility begins at the research stage,” he says. “Researchers have a responsibility for trying to mitigate such problems but also to make users aware of the potential pitfalls — whether a technology is used the way it was intended or, as so often happens, far beyond the initial intentions.”

Recommendations for Researchers

The report recommends several ways to instill social and ethical concerns even in early-stage research.

Government funding agencies, which finance a lot of computer research, can insist that all proposals address such potential risks. Stanford HAI’s funding process requires exactly that kind of ethical and social review. Likewise, professional societies and academic journals can insist that any new published research includes a thorough discussion of potential problems.

More broadly, the report says, research institutions should give computer scientists access to experts from other fields who can offer a broader perspective on potential problems.

“Until relatively recently, many researchers and observers considered computing technologies to be value neutral,” the report notes. “Few, if any, are. The design of new computing technologies … is always imprinted with the spectrum of values considered by the designer, which may not be broad enough to ensure a particular technology meets the needs of some stakeholders.”

“The goal of ethical grounding in technology isn’t to survey every single problem that could possibly occur,’’ Hennessy says. “It’s to make you aware of these issues so that when you encounter a circumstance where there is a potential ethical trade-off, you’ll be cognizant of it and deal with it.”

The NAS panel that produced the report was chaired by Barbara Grosz, a computer scientist at Harvard who is also an honorary member of Stanford’s Institute for Human-Centered Artificial Intelligence. In addition to Hennessy, other panel members included Mark Ackerman, professor human-computer interaction at the University of Michigan; Steve Bellovin, professor of computer science at Columbia University; David Danks,  professor of philosophy and data science at the University of California at San Diego; Mariano-Florentino Cuellar, president of the Carnegie Endowment for International Peace; Megan Finn of the University of Washington; Mary Gray of Microsoft Research; Ayanna Howard of Ohio State University; Jon Kleinberg of Cornell University; James Manyika of the McKinsey Global Institute; James Mickens of Harvard University; and Amanda Stent of Colby College.

Stanford HAI’s mission is to advance AI research, education, policy and practice to improve the human condition. Learn more.

Share
Link copied to clipboard!
Contributor(s)
Edmund L. Andrews

Related News

AI Challenges Core Assumptions in Education
Shana Lynch
Feb 19, 2026
News

We need to rethink student assessment, AI literacy, and technology’s usefulness, according to experts at the recent AI+Education Summit.

News

AI Challenges Core Assumptions in Education

Shana Lynch
Education, SkillsGenerative AIPrivacy, Safety, SecurityFeb 19

We need to rethink student assessment, AI literacy, and technology’s usefulness, according to experts at the recent AI+Education Summit.

Stanford HAI and Swiss National AI Institute Form Alliance to Advance Open, Human-Centered AI
Jan 22, 2026
Announcement
Your browser does not support the video tag.

Stanford, ETH Zurich, and EPFL will develop open-source foundation models that prioritize societal values over commercial interests, strengthening academia's role in shaping AI's future.

Announcement
Your browser does not support the video tag.

Stanford HAI and Swiss National AI Institute Form Alliance to Advance Open, Human-Centered AI

Education, SkillsJan 22

Stanford, ETH Zurich, and EPFL will develop open-source foundation models that prioritize societal values over commercial interests, strengthening academia's role in shaping AI's future.

Citi Begins Retraining 175,000 Employees in Working with AI: ‘Great Prompting Versus Basic Prompting to Generate Impactful Results’
Fortune
Oct 01, 2025
Media Mention

Stanford HAI offers executive education programs for human-centered AI adoption in business practices.

Media Mention
Your browser does not support the video tag.

Citi Begins Retraining 175,000 Employees in Working with AI: ‘Great Prompting Versus Basic Prompting to Generate Impactful Results’

Fortune
Education, SkillsOct 01

Stanford HAI offers executive education programs for human-centered AI adoption in business practices.