Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Lesson 13 — Ethics, Fairness & Responsible AI

(privacy, bias, explainability, and “should we deploy?” thinking)

Why this matters (motivation)

Analytics and AI are not only technical tools—they influence people. Even “simple” models can affect:

Responsible AI means:


Part A — What “responsible AI” means in this course

Two useful frameworks (light use)

We use these as checklists, not as heavy policy documents.


Part B — Privacy and data governance (practical)

What counts as sensitive in typical student projects?

Examples:

In this course:

Privacy basics to communicate in a report


Part C — Fairness and bias: where problems enter

Bias can enter through multiple channels:

1) Target label problem

What is the label measuring?

2) Feature and proxy problem

Some features act as proxies for sensitive attributes:

3) Sampling problem

Your dataset may not represent the population:

4) Measurement problem

Errors may differ across groups:

5) Objective problem

Optimizing overall accuracy can produce unequal harm. Example: a model may reduce errors for majority groups while increasing errors for minority groups.


Part D — Minimal fairness checks (what we can do at this level)

We keep fairness checks simple and transparent.

Step 1: Choose a grouping variable (if available and appropriate)

Examples:

Step 2: Compare performance by group

For classification (e.g., churn yes/no), compare:

For continuous outcomes (regression/forecasting), compare:

Step 3: Interpret carefully

A difference does not automatically prove discrimination. But it is a red flag that requires explanation and monitoring.


Part E — Explainability: what it is and what it isn’t

Many real stakeholders ask: “Why did the model make this prediction?”

Global vs local explainability (simple distinction)

Practical role in your capstone

You may use explainability to:

But always include a caveat:


Part F — “Should we deploy?” (a decision rubric)

Even if you are not actually deploying systems, this is a useful professional habit.


Mini case discussion (in class)

Choose one case:

  1. Credit scoring

  2. Hiring screening

  3. Pricing / eligibility

  4. Student risk prediction (education analytics)

  5. Platform moderation or fraud detection

For the chosen case:


Mini-lab (Google Colab)

In-class checkpoints (privacy & documentation)

  1. Write a short “Data Governance Note” for your capstone dataset:

    • source, access date, unit of observation

    • what personal data exists (if any)

    • storage/access plan

    • limitations

In-class checkpoints (fairness checks)

  1. Choose a grouping variable (if available).

  2. Compute and compare at least one metric by group:

    • classification: accuracy + FN rate (or FP rate)

    • forecasting/regression: MAE by group or by time period

  3. Write 5–7 lines interpreting results and stating one risk.

In-class checkpoints (explainability — concept + demo)

  1. Use a simple explainability tool (or a simplified feature importance):

    • identify top features driving predictions

  2. Identify at least one proxy risk or surprising feature, and describe what you would investigate next.

In-class checkpoints (capstone checklist)

  1. Draft your Responsible AI Checklist for the capstone (see below).

Submission (after class)


Capstone Responsible AI Checklist (required)


AI check (meta)


Review questions (quiz / reflection)

  1. Give two ways bias can enter a model pipeline.

  2. Why can an accurate model still be harmful?

  3. What is the difference between explaining the model and explaining the real world?

  4. Name one fairness check you can do at a basic level in this course.