Blame the Bot? Evaluating Racial Bias from Humans vs AI

Image from Unsplash

Summary

  • This project explores how people evaluate racial bias differently when it originates from a human versus an AI system.
  • It examines how bias explicitness (blatant vs. subtle) and perceiver race (White vs. non-White) influence emotional reactions and blame judgments.
  • Key finding: People hold humans more accountable than AI for biased behavior, but shift blame toward AI trainers—reflecting awareness of human responsibility behind automation.

Challenge

  • Identifying relevant users: Recruited a racially diverse sample to evaluate bias from both human and AI agents across conditions.

  • Establishing user-centered relevance: The study responds to growing public concern about algorithmic bias and accountability, especially among historically marginalized groups.

  • Design realism: Creating vignettes that felt authentic for both human and AI perpetrators without evoking suspicion or confusion.

  • Balancing subtlety: Ensuring “subtle bias” scenarios were recognizable but not too obvious to participants.

  • Measuring moral attribution: Capturing nuanced blame at multiple levels (perpetrator, trainer, society) without inflating survey length.

Solutions to the challenge

  • Developed paired Human vs. AI vignettes with identical wording except for agent identity (“Tod Smith” vs. “TodSMith”).
  • Used pretesting and comprehension checks to validate clarity and emotional impact.
  • Applied linear mixed-effects models (REML) to account for individual variability in emotional and moral responses.

Goal

To understand how people attribute intentionality, accountability, and moral blame when encountering bias from human versus AI agents—and how these judgments differ across racial groups and levels of bias explicitness.

Why it matters

  • Offers insight into how algorithmic bias is morally interpreted by diverse audiences.
  • Informs AI governance and communication strategies, highlighting public intuitions about fairness and accountability.
  • Contributes to the broader dialogue on ethical AI development and social trust in automation.

Research Process

Stage 1: Study design

  • Participants: 224 individuals (52% female; 112 White, 112 non-White) recruited via Prolific.
  • Experimental design: 2 × 2 between-subjects design crossing:
    • Perpetrator type: Human (“Tod Smith”) vs. AI (“TodSMith”)
    • Bias explicitness: Blatant (explicit racial slurs) vs. Subtle (neutral but unfair justification)
  • Vignettes depicted biased hiring decisions across professional contexts.

Study procedure diagram

Stage 2: Measures

  • Dependent variables:
    • Affect, operationalized using the Profile of Mood States (POMS)
    • Perceived racism in the vignette scenario
    • Blame attribution across three levels: perpetrator, AI trainer, and society
    • AI familiarity (measured only in the AI condition)

Stage 3: Analysis

  • Employed ANOVA to test main and interaction effects of agent type, bias explicitness, and perceiver race.

Results

  • Blatant bias evoked stronger emotional upset and higher racism ratings than subtle bias (p < .001).
  • Humans were blamed more than AI for biased actions, especially when bias was blatant (p < .01).
  • AI trainers were blamed more when bias originated from an AI agent (p < .001).
  • No significant interaction of perceiver race on upset or blame attribution.
  • Non-White participants reported greater resentment than White participants (p < .01).

Perceived racism plot



Blame attribution plot 1



Blame attribution plot 2



Blame attribution plot 3

Reflection

This study demonstrates that people differentiate between moral agency and technical causation—seeing AI as less blameworthy but still linked to human responsibility. The findings highlight how public understanding of AI bias intertwines with moral psychology, revealing both empathy and detachment toward algorithmic agents.

Future directions include expanding to cross-cultural samples, manipulating intentionality cues (e.g., self-learning AI vs. human-programmed systems), and examining how outcome severity moderates moral evaluation.

Programming & Tools Summary

Stage Programming Language/Software Key Libraries / Methods
Data Collection Qualtrics
Data Cleaning Python Pandas, NumPy, regex
Statistical Analysis R Base R, lm4
Visualization R ggplot2
Lee Qianqian Cui
Lee Qianqian Cui
Ph.D. Researcher in Social Psychology