Stanford
University
  • Stanford Home
  • Maps & Directions
  • Search Stanford
  • Emergency Info
  • Terms of Use
  • Privacy
  • Copyright
  • Trademarks
  • Non-Discrimination
  • Accessibility
© Stanford University.  Stanford, California 94305.
HAI Weekly Seminar with Kathleen Creel | Stanford HAI
Skip to content
  • About

    • About
    • People
    • Get Involved with HAI
    • Support HAI
    • Subscribe to Email
  • Research

    • Research
    • Fellowship Programs
    • Grants
    • Student Affinity Groups
    • Centers & Labs
    • Research Publications
    • Research Partners
  • Education

    • Education
    • Executive and Professional Education
    • Government and Policymakers
    • K-12
    • Stanford Students
  • Policy

    • Policy
    • Policy Publications
    • Policymaker Education
    • Student Opportunities
  • AI Index

    • AI Index
    • AI Index Report
    • Global Vibrancy Tool
    • People
  • News
  • Events
  • Industry
  • Centers & Labs
Navigate
  • About
  • Events
  • Careers
  • Search
Participate
  • Get Involved
  • Support HAI
  • Contact Us

Stay Up To Date

Get the latest news, advances in research, policy work, and education program updates from HAI in your inbox weekly.

Sign Up For Latest News

Your browser does not support the video tag.
eventSeminar

HAI Weekly Seminar with Kathleen Creel

Status
Past
Date
Wednesday, February 24, 2021 10:00 AM - 11:00 AM PST/PDT
Topics
Ethics, Equity, Inclusion

The Algorithmic Leviathan: Arbitrariness, Fairness, and Opportunity in Algorithmic Decision-Making Systems

Automated decision-making systems implemented in public life are typically highly standardized. One algorithmic decision-making system can replace or influence thousands of human deciders. Each of the humans so replaced had their own decision-making criteria: some good, some bad, and some merely arbitrary. Decision-making based on arbitrary criteria is legal in some contexts, such as employment, and not in others, such as criminal sentencing. Where no other right provides a guarantee of non-arbitrary decision-making, is arbitrariness of moral concern?

An isolated arbitrary decision need not morally wrong the individual whom it misclassifies. However, if the same algorithms produced by the same companies are uniformly applied across wide swathes of a public sphere, be that hiring or lending, the same people could be consistently excluded from employment, loans, or other sectors of civil society. This harm persists even when the automated decision-making systems are “fair” on standard metrics of fairness.  We argue that arbitrariness at scale is morally and should be legally problematic. The heart of this moral issue relates to domination and a lack of sufficient opportunity for autonomy.  It relates in interesting ways to the moral wrong of discrimination. We propose technically informed solutions that can lessen the impact of algorithms at scale and so mitigate or avoid the moral harm we identify.  

Speaker
Kathleen Creel
HAI Network Affiliate; Assistant Professor of Philosophy and Computer Science, Northeastern University

Watch Event Recording

Share
Link copied to clipboard!
Event Contact
Celia Clark
celia.clark@stanford.edu
More from HAI and SDS seminars
  • Hari Subramonyam | Learning by Creating: A Human-Centered Vision for AI in Education
    SeminarMar 11, 202612:00 PM - 1:15 PM
    March
    11
    2026

Related Events

Gaidi Faraj, Lofred Madzou | Nurturing Africa’s AI Leaders through Math Olympiad
SeminarFeb 25, 202612:00 PM - 1:15 PM
February
25
2026

The African Olympiad Academy is a world-class high school dedicated to training Africa’s most promising students in mathematics, science, and artificial intelligence through olympiad-based pedagogy.

Seminar

Gaidi Faraj, Lofred Madzou | Nurturing Africa’s AI Leaders through Math Olympiad

Feb 25, 202612:00 PM - 1:15 PM

The African Olympiad Academy is a world-class high school dedicated to training Africa’s most promising students in mathematics, science, and artificial intelligence through olympiad-based pedagogy.

Dan Iancu & Antonio Skillicorn | Interpretable Machine Learning and Mixed Datasets for Predicting Child Labor in Ghana’s Cocoa Sector
SeminarMar 18, 202612:00 PM - 1:15 PM
March
18
2026

Child labor remains prevalent in Ghana’s cocoa sector and is associated with adverse educational and health outcomes for children.

Seminar

Dan Iancu & Antonio Skillicorn | Interpretable Machine Learning and Mixed Datasets for Predicting Child Labor in Ghana’s Cocoa Sector

Mar 18, 202612:00 PM - 1:15 PM

Child labor remains prevalent in Ghana’s cocoa sector and is associated with adverse educational and health outcomes for children.