Course:AI & Tools Literacy/
Lesson

AI reflects the biases of its training data and the society that created it. Using AI responsibly means understanding these limitations and making conscious choices about when and how to use it.

Where AI bias comes from

Bias in AI isn't about malice, it's a consequence of how models are built. AI models learn from human-created data that contains all the biases of human society:

  • Historical bias: Past hiring data reflects discrimination; medical data underrepresents women and minorities; financial data may reflect redlining
  • Representation bias: More English than other languages, more Western perspectives, more data about majority groups
  • Measurement bias: What we measure reflects our values; how we label data introduces judgment; proxy variables encode hidden biases

Amplification in the model

The training process can amplify biases beyond what exists in the data:

Training data: "Doctors are often male, nurses often female"

AI learns: Strong association between "doctor" and male pronouns,
           Strong association between "nurse" and female pronouns

Result: When asked to complete "The doctor walked into the room,
        ___ examined the patient," AI overwhelmingly chooses "he"
02

Types of bias in AI systems

Representation bias

When certain groups are underrepresented in training data, the AI performs worse for them:

DomainIssueExample
Facial recognitionLess accurate for darker skin tonesHigher false positive rates for Black faces
Speech recognitionTrained mostly on certain accentsPoor performance on non-American accents
Medical diagnosisUnderrepresented in clinical trialsMissed symptoms in women and minorities
Language modelsMore data in dominant languagesLower quality for low-resource languages

Stereotyping and associations

AI reinforces harmful stereotypes, gender stereotypes in profession associations, racial stereotypes in crime-related queries, socioeconomic biases in credit and hiring contexts.

Under the hood
Researchers found GPT-3 associated "brilliant" and "genius" more with male names, while "nurturing" and "empathetic" associated more with female names, patterns learned from decades of human-written text.

Feedback loop bias

When AI predictions influence the world, they create self-reinforcing cycles:

1. AI predicts certain neighborhoods have higher crime risk
2. Police allocate more resources to those neighborhoods
3. More arrests occur in those neighborhoods
4. New data shows even higher crime rates
5. AI becomes more confident in its initial prediction

The AI's prediction became a self-fulfilling prophecy.

03

Real-world impacts of AI bias

Hiring: Amazon's AI hiring tool, trained on 10 years of resumes, learned to penalize resumes containing "women's" (as in "women's chess club captain") because the tech industry historically hired more men. The system was abandoned.

Healthcare: A hospital AI systematically underestimated Black patients' medical needs. It used healthcare costs as a proxy for health needs, but Black patients historically had less access to care (lower costs), so the AI concluded they were healthier.

Criminal justice: COMPAS, a risk assessment tool used in bail and sentencing, showed racial disparities, Black defendants were more likely flagged as high risk, while white defendants with similar profiles were rated lower risk.

04

Ethical considerations in AI use

Transparency

Disclose when content is AI-generated or AI-assisted. Don't pass off AI work as entirely human-created. Be clear about AI capabilities and limitations with users.

Consent

Was training data collected with appropriate consent? Are users aware their inputs might train models? Do people know AI is analyzing their images, voice, or writing?

Accountability

If AI-generated code has a security vulnerability, who is responsible? If biased AI makes discriminatory decisions, who answers for it?

Good to know
Most jurisdictions treat the human deploying AI as responsible, but this is evolving. As an AI user, assume you're accountable for outputs you use or deploy.

Harm reduction

Consider: direct harm (physical, emotional, financial), indirect harm (misinformation, discrimination, privacy violations), systemic harm (job displacement, concentration of power), and environmental harm (energy consumption of large models).

05

When to involve humans

Some decisions should never be fully automated:

  • High-stakes individual decisions: Medical diagnoses, criminal justice, hiring/firing, credit approvals, child welfare
  • Value-laden judgments: Content moderation, resource allocation, research funding
  • Novel situations: AI works on patterns from the past, unprecedented situations require human judgment
06

Responsible AI practices

As an AI user

  • Audit outputs: Check if suggestions seem biased, test with diverse inputs, question whether outputs reinforce stereotypes
  • Document your use: Note when AI assisted your work, keep records of prompts and outputs
  • Stay informed: Keep up with AI capabilities, limitations, and ethics developments
  • Know when not to use AI: Don't use it for decisions about individuals' rights, when you can't verify output, or for tasks requiring genuine understanding

When building AI-powered products

  • Test for bias: Include diverse test cases, check performance across demographic groups, use bias detection tools
  • Build in oversight: Design systems with human review points, make AI recommendations (not decisions), allow users to contest outputs
  • Be transparent: Document what your AI does and doesn't do, explain how decisions are made

AI is not a neutral tool. It reflects our society, our biases, and our choices. Using it responsibly means staying aware of its limitations, considering its impacts, and keeping human judgment at the center of important decisions.