Why this dimension matters
The assertion gap. Most organizations claim their AI is fair and explainable. Very few can show you the numbers. Responsibility is the dimension where the gap between policy language (“we are committed to fairness”) and operational reality (“our last release changed the bias delta on the loan-approval model from 3.2pp to 5.8pp and we caught it in the staging gate”) is widest. Closing that gap is what distinguishes a governance program from a governance brochure.
The regulatory horizon. The EU AI Act (Articles 10, 13, 14) requires that high-risk AI systems be trained on representative data, offer meaningful transparency to users, and support human oversight. The NIST AI RMF names fairness, explainability, and accountability as four of its seven trustworthy AI characteristics. ISO/IEC 23894 requires documented evidence of bias assessment. These are not aspirational statements — they are auditable obligations. A Responsibility metric program is how you meet them on a schedule.
The downstream blast radius. Responsibility failures do not stay inside the model. A biased hiring screen produces a class action. An unexplainable credit decision produces a regulatory enforcement letter. An unfair benefits-eligibility model produces front-page news. The cost of measuring is always smaller than the cost of discovering.
What good looks like
A mature Responsibility measurement program has five properties:
- Every high-risk model has a named fairness owner and an approved set of protected attributes, documented in the model card.
- Bias is measured pre-release and post-release on the same evaluation suite so drift is visible.
- Explainability is tiered (global, local, counterfactual) and the tier required is chosen by use case risk, not by developer preference.
- Thresholds are published and enforced by a gate — a model that exceeds the bias-delta threshold cannot ship without an approved exception.
- Results are reported to the governance board on the same cadence as financial metrics.
Core metrics
Metric 1: Bias delta across protected groups
Definition. The maximum difference in a chosen performance metric (accuracy, true positive rate, false positive rate, selection rate, or calibration error) across the protected groups defined for the use case, measured on a fixed evaluation dataset.
Formula. bias_delta = max(metric[group_i]) − min(metric[group_j]) for all groups i, j in the protected attribute set.
Cadence. Measured on every model release candidate and re-measured monthly on production traffic samples.
Owner. Model owner, with review by the Responsible AI lead.
Data source. A versioned fairness evaluation dataset plus a representative production traffic sample drawn under the model’s data-use policy.
Metric 2: Explainability coverage
Definition. The percentage of in-scope model decisions for which a documented explanation of the required tier (global, local, or counterfactual) is available on demand within the service-level time budget.
Formula. explainability_coverage = (decisions_with_available_explanation / total_decisions_in_window) × 100.
Cadence. Measured continuously; reported weekly.
Owner. Platform engineering, with policy input from the Responsible AI lead.
Tier selection. Use-case risk tier drives the required explanation tier. Low-risk models may only require a global feature-importance report. High-risk models (credit, hiring, healthcare, benefits) require per-decision local explanations, and counterfactual explanations (“what would have to change for the decision to flip”) are required wherever a regulation grants a right to contest.
Metric 3: Fairness composite
Definition. A weighted index combining bias delta, disparate impact ratio, and calibration error into a single 0–100 score, where 100 is the stated fairness target and scores below the alert threshold trigger a review.
Formula. fairness_composite = w1 × normalize(bias_delta) + w2 × normalize(disparate_impact_gap) + w3 × normalize(calibration_error) with weights documented per model.
Cadence. Monthly, published on the trust scorecard.
Owner. Responsible AI lead.
The composite exists because no single fairness metric captures every harm. The four-fifths rule catches selection-rate disparity but misses calibration drift; equalized odds catches error-rate disparity but misses the four-fifths rule. The composite forces the team to pick the right combination for the use case and then hold the combined line.
How to measure — step by step
- Inventory the protected attributes. For each high-risk model, document which attributes are in scope (age, sex, race, disability, national origin, and any jurisdiction-specific class). Record the lawful basis for collecting or inferring each attribute.
- Build the evaluation dataset. Create a versioned dataset with sufficient representation for every protected group. Under-represented groups produce unstable metrics — if a group has fewer than 100 examples, the confidence interval on bias delta will dominate any signal.
- Pick the primary metric. Accuracy parity is rarely the right choice. Selection rate parity is appropriate for allocation decisions (loans, hiring, benefits). Error-rate parity is appropriate for diagnostic decisions. Calibration parity is appropriate for risk-scoring decisions.
- Measure on the release candidate. Run the evaluation before release. Record bias delta, disparate impact ratio, and calibration error per group.
- Set the gate. Publish the threshold ahead of time. The gate rejects the release or triggers an exception workflow if any metric exceeds its threshold.
- Measure on production traffic. Sample live traffic monthly, re-evaluate, and compare to the pre-release baseline. Drift above 20% of the original delta is a material change and triggers re-review.
- Report to the board. The fairness composite joins the value, reliability, and compliance metrics on the single trust scorecard.
Targets and thresholds
These are defaults, not requirements — every program must set its own thresholds in line with its risk appetite and regulatory environment.
- Bias delta (selection rate). Disparate impact ratio must remain above 0.80 (the four-fifths rule from the EEOC Uniform Guidelines). Alert threshold 0.85.
- Bias delta (error rate). Maximum per-group difference under 5 percentage points for high-risk models. Alert threshold 3 percentage points.
- Explainability coverage. 99% for high-risk, 95% for medium-risk, 90% for low-risk.
- Fairness composite. Minimum 80 out of 100 for production release; scores between 80 and 90 require a documented mitigation plan.
Common pitfalls
Measuring one metric and calling it fairness. A model can pass the four-fifths rule and still be unfair on calibration. Use the composite.
Evaluating on training data. Bias measured on training data is meaningless for production risk. Always measure on a held-out, production-representative set.
Hiding behind aggregates. If a protected attribute has multiple values (race, for example), reporting only “the maximum group delta” hides which group is being harmed. Publish per-group metrics, not just the max.
Treating explainability as a UX feature. Explainability is a decision-integrity property. If you cannot explain a decision, you cannot defend it in court, cannot remediate it for the affected person, and cannot learn from its failures.
Exception without expiry. Every approved exception to a fairness threshold must have a named owner, a remediation plan, and an expiry date — or the exception becomes the new standard.
Related articles
M3.4Advanced Ethics Architecture M3.4AI Risk Governance at Enterprise Scale M1.3Governance Pillar Domains — Strategy, Ethics, and Compliance M4.3NIST AI RMF Implementation at Enterprise Scale M4.3ISO 42001 Alignment and AI Management System Certification