Lab 14: Responsible AI Audit

Time: 50 minutes | Level: Architect | Docker: docker run -it --rm zchencow/innozverse-ai:latest bash

Overview

Responsible AI requires systematic auditing for fairness, explainability, and accountability. This lab covers fairness metrics, bias detection, SHAP/LIME explainability, model cards, AI impact assessments, and building an automated audit trail.

Architecture

┌──────────────────────────────────────────────────────────────┐
│                Responsible AI Audit Framework                 │
├──────────────────────────────────────────────────────────────┤
│  FAIRNESS ASSESSMENT         │  EXPLAINABILITY               │
│  ────────────────────        │  ──────────────────           │
│  Demographic parity          │  SHAP (global + local)        │
│  Equalized odds              │  LIME (local)                 │
│  Calibration                 │  Feature importance           │
│  Disparate impact (80% rule) │  Counterfactuals              │
├──────────────────────────────┴──────────────────────────────┤
│  MODEL CARD  │  IMPACT ASSESSMENT  │  AUDIT TRAIL           │
│  Capabilities│  Stakeholder harm   │  Immutable logs         │
│  Limitations │  Risk register      │  Decision records       │
│  Metrics     │  Mitigation plan    │  Version history        │
└──────────────────────────────────────────────────────────────┘

Step 1: Fairness Metrics Taxonomy

No single fairness metric covers all situations. Choose based on your use case and legal context.

Group Fairness Metrics:

Metric
Definition
Formula
Use Case

Demographic parity

Equal positive rates across groups

P(Ŷ=1|A=0) = P(Ŷ=1|A=1)

Hiring, loans

Equalized odds

Equal TPR and FPR across groups

TPR₀=TPR₁ and FPR₀=FPR₁

Criminal justice

Equal opportunity

Equal TPR across groups

TPR₀=TPR₁

Medical diagnosis

Calibration

Predicted prob equals actual prob

P(Y=1|Ŷ=p,A=a) = p for all a

Risk scoring

Individual fairness

Similar people treated similarly

f(x)≈f(x') if x≈x'

General

Disparate Impact (Legal "80% Rule"):

💡 Fairness is mathematically impossible to optimize all metrics simultaneously (Chouldechova, 2017). Choose the metric that aligns with your legal and ethical framework.


Step 2: Bias Detection

Sources of Bias:

Source
Description
Example

Historical bias

Training data reflects past discrimination

Credit data: women historically denied loans

Representation bias

Training data underrepresents some groups

Facial recognition trained mostly on white faces

Measurement bias

Different measurement accuracy across groups

Medical sensors less accurate on darker skin

Aggregation bias

One model for diverse subgroups

Diabetes model ignores ethnic variation

Deployment bias

System used in different context than designed

Designed for urban, deployed in rural

Simpson's Paradox:

Bias Audit Checklist:


Step 3: SHAP Explainability

SHAP (SHapley Additive exPlanations) provides consistent, theoretically-grounded feature importance.

SHAP Values:

SHAP Visualizations:

Plot
Shows
Use

Beeswarm

Feature impact distribution

Global: which features matter most

Waterfall

Single prediction breakdown

Local: why this specific decision

Bar chart

Mean

SHAP

Dependence

Feature value vs SHAP value

Feature interaction detection

Force plot

Single prediction visualization

User-facing explanation

SHAP for Compliance:


Step 4: LIME Explainability

LIME (Local Interpretable Model-Agnostic Explanations) explains any model locally.

LIME Process:

LIME vs SHAP:

Dimension
LIME
SHAP

Scope

Local only

Local + Global

Speed

Fast

Can be slow

Consistency

Stochastic (varies per run)

Deterministic

Theoretical basis

Heuristic

Game theory (Shapley values)

Model support

Any

Any (but efficient for tree models)

Best for

Quick local explanations

Rigorous global + local analysis


Step 5: Model Cards

Model cards (Google, 2019) document model behavior, limitations, and intended use.

Model Card Template:


Step 6: AI Impact Assessment

Structured assessment of potential harms before deploying AI systems.

Impact Assessment Framework:


Step 7: Audit Trail Architecture

Immutable Audit Log Requirements:

Audit Trail Storage:

Linking Predictions to Outcomes:


Step 8: Capstone — Fairness Metrics Calculator

📸 Verified Output:


Summary

Concept
Key Points

Fairness Metrics

Demographic parity, equalized odds, disparate impact (80% rule)

Simpson's Paradox

Always evaluate fairness per subgroup, not just aggregate

SHAP

Consistent feature importance; sum = model output; local + global

LIME

Local explanations; black-box compatible; stochastic

Model Cards

Document: performance per group, limitations, intended use, owners

AI Impact Assessment

Structured harm identification before deployment

Audit Trail

Append-only, immutable; link predictions to outcomes; 7-year retention

Next Lab: Lab 15: LLM Fine-Tuning Infrastructure →

Last updated