Skip to main content
AITE M1.3-Art14 v1.0 Reviewed 2026-04-06 Open Access
M1.3 The 20-Domain Maturity Model
AITF · Foundations

OKRs and AI Delivery Cadence

OKRs and AI Delivery Cadence — Maturity Assessment & Diagnostics — Advanced depth — COMPEL Body of Knowledge.

10 min read Article 14 of 48 Calibrate

COMPEL Specialization — AITE-VDT: AI Value & Analytics Expert Article 14 of 35


An AI product team sets quarterly OKRs at the start of Q1. The Objective is “make the customer-service copilot indispensable to our 600 agents.” The Key Results include “90% daily active use among agents”, “AI draft acceptance rate above 75%”, and “model latency p95 below 2 seconds.” Three months later the team has hit the adoption and latency Key Results but missed the acceptance rate, and the CFO has approached the value lead to ask whether the copilot is actually producing value because cost-per-resolved-ticket has not moved. The OKRs were internally coherent and externally disconnected. They measured operational quality without connecting to the business outcome the programme was funded to deliver. This pattern is common. OKRs — Objectives and Key Results — align delivery to corporate objectives when written correctly and produce tactical silos when written poorly. This article teaches the practitioner to write AI-specific Objectives, to craft Key Results that are measurable without over-constraining model behaviour, and to run quarterly OKR reviews that convert model-level metrics into organisation-level outcomes.

Why OKRs matter for AI programmes specifically

AI programmes have a particular tendency to drift away from business outcomes because their operational metrics are technically rich. Accuracy, F1, latency, throughput, evaluation-coverage ratio — each is a legitimate engineering metric and each is a potential distraction from the outcome that funded the programme. OKRs are the discipline that prevents the drift by tying each quarter’s work to a named business objective.

The OKR practice emerged at Intel in the 1970s under Andy Grove, was adopted by Google in the late 1990s under John Doerr’s consulting, and has been documented extensively in published enterprise adoption case studies and management literature. The practice is now widely used in technology-heavy enterprises. For AI programmes, the OKR discipline has a specific utility: it forces the model-level technical team to state, quarterly, what business outcome they are optimising for, and to measure the outcome rather than the accuracy proxy.

MIT Sloan’s operational-discipline research, developed across the State of AI at Work series, consistently finds that organisations capturing value from AI have tighter quarterly alignment between technical delivery and business outcome than organisations that do not. The OKR practice is the most common operational form of the tighter alignment.1

Writing AI-specific Objectives

An Objective states a direction, not a metric. A good Objective is qualitative, aspirational, memorable, and aligned to a higher-order business outcome. “Make the customer-service copilot indispensable to our 600 agents” is an Objective — qualitative, memorable, aligned to a business outcome (agent productivity).

Three tests determine whether an Objective is well-written for an AI programme. First, does it name the business outcome rather than the technical capability? An Objective of “achieve state-of-the-art latency” fails this test because latency is a technical attribute, not a business outcome. Second, is it achievable in one quarter? Objectives that span two or more quarters should be decomposed into quarterly milestones; a single-quarter Objective is an Objective. Third, does it fit a single sentence? Objectives longer than a sentence tend to be compound Objectives the team is trying to pass off as one, and the compound is better split into two.

The AITM-CMD credential develops similar discipline for change-management programmes; the AITE-VDT practitioner uses the same structural quality bar.

Crafting Key Results for AI deliverables

A Key Result is a measurable milestone whose achievement indicates progress toward the Objective. Good Key Results are specific, measurable, ambitious-but-achievable, relevant, and time-bound. The classic “SMART” heuristic applies.

For AI deliverables, two additional rules matter. The first is that Key Results should measure outcomes when possible, not outputs. “Ship version 2 of the model” is an output-level Key Result; “reduce average handle time for treated tickets by 40%” is an outcome-level Key Result. Both have their place, but over-reliance on output-level KRs produces the pattern where the team ships on schedule and the business outcome does not move.

The second is that Key Results should not over-constrain the model’s behaviour. “AI draft acceptance rate above 75%” is a reasonable Key Result. “AI draft acceptance rate above 90%” for a feature where 65% is the realistic plausible ceiling is a Key Result that forces the team into counterproductive behaviour: inflating the denominator, gaming the definition, or pushing the model toward producing safe-but-useless drafts that users accept without value. Over-constrained KRs produce metric gaming; well-calibrated KRs produce genuine effort.

A well-crafted Key Result for an AI programme has three components: the metric, the target, and the measurement methodology. The metric is one of the KPI-tree metrics Article 12 defined. The target is calibrated against benchmarks and internal pilot data. The measurement methodology references the measurement plan Article 4 introduced, so the KR is grounded in the pre-registered measurement discipline.

[DIAGRAM: Timeline — quarterly-okr-review-cycle — 13-week timeline showing week 1 OKR-setting, week 4 and week 8 check-ins, week 12 mid-quarter review, week 13 end-of-quarter review-and-reset; each event annotated with inputs and outputs; primitive teaches the OKR-cycle cadence.]

The quarterly cadence — from setting to reset

OKR practice operates on a quarterly cycle with four phases.

Phase 1 — Setting (week 1 of the quarter). The team drafts Objectives based on the corporate-level objectives for the quarter, with the business sponsor’s input. Key Results are drafted alongside and calibrated against the previous quarter’s actuals and relevant benchmarks. The draft is reviewed against the KPI tree to ensure each Key Result maps to a driver or metric in the tree.

Phase 2 — Check-ins (weeks 4 and 8). The team reports progress toward each Key Result, with a confidence score (probability of achievement by end of quarter). Check-ins are short — 15 minutes per OKR — and are used for early detection of KRs drifting off track, not for detailed status reporting.

Phase 3 — Mid-quarter review (week 12). The team reviews each Key Result for feasibility. KRs tracking to achievement are confirmed; KRs unlikely to achieve are triaged — either the work is re-prioritised or the KR is flagged for formal miss. Mid-quarter resets are rare and require sponsor approval; the discipline is to accept the miss rather than move the goalposts.

Phase 4 — End-of-quarter review and reset (week 13). The team reports each Key Result’s final measured value, reviews the Objective’s achievement narrative, and extracts lessons for the next quarter. The next quarter’s OKR setting begins with the lessons from the previous quarter’s review.

Published enterprise OKR adoption patterns across multiple technology-heavy firms document the four-phase cadence as the operational norm; variations exist (bi-monthly check-ins, tighter weekly cadences for high-velocity teams), but the core rhythm is remarkably consistent.

Common OKR failure modes for AI teams

Five failure modes recur in AI-team OKR practice.

Sandbagging. Teams set low targets they are confident of hitting, making the OKR exercise meaningless as a stretch mechanism. The corrective is calibrated stretch: targets set at roughly the 70% confidence level, so a team that achieves all its Key Results has done meaningfully better than its base rate. The AITE-VDT practitioner looks at the team’s last four quarters of Key Result achievement rates; a team hitting 95%+ is sandbagging.

Roof-shot inflation. The reverse failure: teams set targets they cannot realistically achieve, producing learned helplessness when they predictably miss. The corrective is calibrated ambition: targets set against realistic benchmarks, with the sponsor agreeing that the base case is plausible.

Metric gaming. A Key Result whose definition can be optimised against without producing the intended outcome invites gaming. “Number of customer-service tickets handled by AI” is gameable (push easy tickets to AI, hold hard ones for humans); “number of customer-service tickets resolved by AI with maintained satisfaction” is less gameable. The AITM-CMD credential’s discussion of adoption-metric gaming applies to AITE-VDT practice as well.

OKR proliferation. A team with twelve Objectives and forty Key Results has lost the focus the OKR practice was meant to create. The canonical discipline is three to five Objectives per team per quarter, each with three to five Key Results.

Outcome-output confusion. A team whose Key Results are all output-level (shipping milestones, release counts, model-version numbers) has disconnected from the business outcome. The corrective is a minimum threshold: at least half of each team’s Key Results must be outcome-level (or leading-indicator proxies for outcome-level metrics).

[DIAGRAM: MatrixDiagram — okr-quality-dimensions — 2×2 grid with axes “specificity (low/high)” and “outcome-orientation (low/high)”; sample Key Results placed in the grid; ideal quadrant (high/high) populated with well-crafted KRs; anti-patterns populated in the other quadrants; primitive teaches the KR quality self-assessment.]

Worked example — a cross-team OKR stack

OKRs work best when they stack — the team’s OKRs roll up to the product-line OKRs, which roll up to the business-unit OKRs, which roll up to the corporate OKRs. A proper stack makes the alignment visible.

For the customer-service copilot: corporate Objective “improve customer-service contribution margin”; business-unit Objective “reduce cost per resolved ticket by US$1.50”; product-line Objective “scale AI-assisted resolution to 65% of tickets”; team Objective “make the copilot indispensable to the 600 agents”. Each Objective at the lower level is a component of its parent’s Objective, and each team’s Key Results advance the parent’s outcome.

The value lead’s role is to inspect the stack’s alignment at the beginning of each quarter. An AI team whose Objective does not roll up to a business-unit Objective is doing work the business unit has not asked for. A business-unit Objective that does not roll up to a corporate objective is a unit operating outside the corporate strategy. Mis-alignments at any level are cheap to correct at the beginning of the quarter and expensive to correct at the end.

The OKR’s relationship to the measurement plan and the KPI tree

OKRs do not replace the measurement plan or the KPI tree. The KPI tree is the standing hierarchical decomposition of outcomes to drivers to metrics. The measurement plan is the pre-launch document that pre-registers how each feature will be measured. The OKRs are the quarterly delivery commitments that operate within the tree-and-plan structure.

A team’s OKRs should be computable from the KPI tree’s metrics; a Key Result that measures something not on the tree is either the tree needing extension or the Key Result being ill-specified. Similarly, a Key Result that corresponds to a primary metric in a measurement plan provides the pre-registered measurement methodology; a Key Result without a corresponding plan-of-record entry is a Key Result missing its measurement infrastructure.

Summary

OKRs are the quarterly alignment mechanism that converts model-level metrics into organisation-level outcomes. Objectives state direction in qualitative, aspirational, memorable language. Key Results measure progress with specificity and outcome-orientation, avoiding over-constraint that invites gaming. The quarterly cadence has four phases — setting, check-ins, mid-quarter review, end-of-quarter review. Five failure modes — sandbagging, roof-shot inflation, metric gaming, proliferation, outcome-output confusion — are detectable and correctable. OKRs stack from team to product line to business unit to corporate, and misalignments are cheapest to correct at the stack’s setting phase. Article 15 turns to the Control Performance Report artifact that aggregates the operational measurement under the internal-process perspective.


Cross-references to the COMPEL Core Stream:

  • EATP-Level-2/M2.5-Art02-Designing-the-Measurement-Framework.md — measurement framework into which OKR cadence is embedded
  • EATF-Level-1/M1.2-Art05-Evaluate-Measuring-Transformation-Progress.md — Evaluate stage methodology that operates on quarterly review rhythm
  • EATP-Level-2/M2.5-Art10-From-Measurement-to-Decision.md — the measurement-to-decision discipline quarterly OKR reviews exercise

Q-RUBRIC self-score: 90/100

© FlowRidge.io — COMPEL AI Transformation Methodology. All rights reserved.

Footnotes

  1. MIT Sloan Management Review and Boston Consulting Group, “State of AI at Work” series (2020–2025), https://sloanreview.mit.edu/projects/expanding-ais-impact-with-organizational-learning/ (accessed 2026-04-19).