COMPEL Specialization — AITM-AAG: Agentic AI Governance Associate Article 5 of 14
Definition. Human oversight, in the EU AI Act Article 14 sense, is the set of measures designed and built into a high-risk AI system so that a human can monitor it, interpret its outputs, decide whether to use them, intervene, and — if necessary — halt it. Article 14(4) enumerates the specific capabilities the oversight arrangement must give the human. Oversight theatre is the failure mode in which the oversight mechanism exists on paper but does not, in practice, confer any of those capabilities.
Article 14 is not optional. For a high-risk system under the Act (Articles 6(2) and Annex III, among others, covered in the AITB-RCM credential), the provider must design the system so that effective oversight is possible, and the deployer must assign competent oversight personnel. The specialist’s job is to turn both obligations into a working design: what the oversight operator sees, when, what they can do, and how the system ensures they can actually do it.
What Article 14 requires
Article 14(1) requires that high-risk AI systems be designed and developed — including with appropriate human-machine interface tools — such that they can be effectively overseen by natural persons during the period in which the system is in use. Article 14(4) then lists the operational capabilities that the oversight must confer on those natural persons. The specialist should read the article text in full; a paraphrase is reproduced here for working reference.
| Article 14(4) requirement | Plain language |
|---|---|
| (a) properly understand the relevant capacities and limitations of the system | Operators know what the system can and cannot do. |
| (b) remain aware of the possible tendency of automatically relying or over-relying on the output | Operators are trained against automation bias. |
| (c) correctly interpret the system’s output | Operators can read outputs in context. |
| (d) decide, in any particular situation, not to use the system, or to disregard, override or reverse its output | Operators have genuine discretion and authority. |
| (e) intervene on the operation of the system or interrupt the system through a “stop” button or a similar procedure | Operators can halt the system. |
For agentic systems, every one of those requirements carries added weight. An autonomous executor (Level 4) that runs for hours between operator views has to earn requirement (e) with engineering effort, not just a policy assertion. An agent whose tool calls operate on external systems has to make requirement (c) possible by emitting interpretable logs, not just raw traces.
Source for Article 14 text: Regulation (EU) 2024/1689, https://eur-lex.europa.eu/eli/reg/2024/1689/oj.
Four oversight modes
The Article 14 requirements translate, in practical design, into four oversight modes. A given agent uses one, several, or all four.
Pre-authorisation oversight
The operator reviews and approves a proposed action before it executes. The mode is appropriate for high-consequence actions (money movement, external communication, data export, production write) where post-hoc detection is insufficient. Pre-authorisation is the lowest-automation mode — every qualifying action waits for human action — and therefore the most expensive in operator time.
Design choices:
- Which action classes require pre-authorisation?
- What information does the operator see to authorise?
- What is the timeout if the operator does not respond?
- Is there a fallback authority (a second operator) for out-of-hours windows?
Runtime intervention
The operator monitors the agent during execution and can interrupt, redirect, or halt. The mode is appropriate for medium-consequence actions where most execution proceeds without oversight but operators retain the ability to stop a running process.
Design choices:
- Is the monitoring synchronous (the operator watches a dashboard) or asynchronous (alerts fire when specified conditions are met)?
- What signals surface to the operator?
- What is the latency from alert to intervention?
- Is the intervention granular (stop this action) or coarse (stop the whole session)?
Post-hoc review
The operator reviews actions after completion. The mode is appropriate for lower-consequence actions where the volume is too high for per-action review but pattern review is necessary.
Design choices:
- What is the sample rate?
- What triggers a deeper review?
- Who has authority to reverse, refund, or escalate based on the review?
- What is the feedback loop into the agent’s prompt, tools, or retirement?
Stop-go decision right
The authority to halt the agent entirely, or to resume it after a pause, rests with a named role. The right is exercised outside normal operation — during incidents, planned change, or policy review. It is different from the runtime intervention mode because it acts on the agent as a whole, not on individual actions.
Design choices:
- Who holds the right — a single named role, a committee, a 24-hour operations line?
- What evidence supports a stop decision?
- What conditions must be satisfied to resume?
- How is the decision recorded?
The oversight-theatre failure mode
The Article 14 language is compliance-friendly. A deployer can write policies that assert each of requirements (a)-(e) without the underlying design or operator capability to support them. The failure is predictable and widespread. Four signs that an oversight regime is theatre rather than substance:
- Operator cannot explain how the agent reaches its output. The agent is a black box to the operator. Requirement (a) is unmet.
- High approval rate with low modification rate. Operators are rubber-stamping. Requirement (d) is nominally present but functionally absent. Automation bias has eaten the oversight.
- The “stop” button exists but is not rehearsed. The button’s wiring, the authority for pressing it, and the recovery procedure are all untested. Requirement (e) works in theory only.
- No measurable oversight activity log. The regulator or auditor asks “how often did operators intervene?” and no number exists. Whatever is happening, it is not being measured.
Oversight theatre is usually unintentional. It emerges from an engineering team that solved the system design problem as stated (“show the operator the output”) without solving the stated problem (“give the operator the ability to decide”). The specialist’s role is to call the distinction early, before the theatrical version ships.
Designing an oversight regime — step by step
The design follows from the authority chain (Article 4) and the autonomy classification (Article 3). The output is a written oversight specification that the engineering team builds and the operations team runs.
- Enumerate the agent’s action types. Group by consequence (reversible vs. irreversible; internal vs. external; reputational; financial; safety).
- Assign each action type to one of the four modes. Irreversible external actions default to pre-authorisation. Reversible internal actions can default to post-hoc review. Boundary cases go to runtime intervention.
- Specify the signal and latency for each mode. What does the operator see? How fast? In what interface?
- Name the operator role and competency. Article 14(5) requires the oversight operator to be a natural person with specific knowledge. The competency is a training and selection decision, not an assumption.
- Rehearse the stop-go right. Fire drills, recorded, with lessons captured.
- Instrument the measurement. Every oversight intervention should be logged with operator identity (by role, not by individual in the public artifact), timestamp, context, and outcome.
The specification is reviewed by the Methodology Lead and, for EU high-risk deployments, is a required input to the conformity-assessment file.
Two real-world anchors for oversight design
EU AI Act Article 14 — primary regulatory text
The regulation itself is the first anchor. The specialist should have the text to hand and should read Article 14 in the context of Articles 16, 26, and Annex III. The consolidated text is available from EUR-Lex at https://eur-lex.europa.eu/eli/reg/2024/1689/oj. The AITB-RCM credential treats the Article 14 obligation in greater depth against classification decisions; this article treats it specifically as an oversight-design input for agents.
UK MHRA AI in medical devices guidance — regulated-sector example
The UK Medicines and Healthcare products Regulatory Agency has published guidance on AI in medical devices that names human oversight requirements in a regulated-health context. The guidance is not a direct EU AI Act source, but the two regulatory traditions are converging, and the MHRA material is useful as a model of how a sectoral regulator translates “effective human oversight” into operational design expectations. The governance analyst reads it for pattern, not for literal application to EU AI Act cases. Source: https://www.gov.uk/government/publications/software-and-artificial-intelligence-ai-as-a-medical-device.
Oversight at each autonomy level
The four modes combine differently by autonomy level.
| Level | Typical mode mix |
|---|---|
| 0 — Assisted | Pre-authorisation (the human prompts) |
| 1 — Advisor | Pre-authorisation + post-hoc review of advisory quality |
| 2 — Bounded executor | Pre-authorisation per action + stop-go |
| 3 — Supervised executor | Pre-authorisation at plan + runtime intervention + stop-go |
| 4 — Autonomous executor | Runtime intervention + post-hoc review + stop-go |
| 5 — Self-directing | All four modes, with defence-in-depth on stop-go |
The rising oversight burden at higher levels is a feature. An organisation that cannot staff or engineer the required regime at a given level should not run agents at that level. The correct response to “we don’t have the oversight capacity for Level 4” is to run the system at Level 3, not to declare Level 4 with paper-only oversight.
Learning outcomes — confirm
A specialist who completes this article should be able to:
- Recite Article 14(4) requirements in operational terms.
- Name the four oversight modes and assign action types to them.
- Identify oversight theatre signals in a described agent deployment.
- Produce an oversight specification for a described Level 3 agent that satisfies Article 14.
Cross-references
EATP-Level-2/M2.4-Art11-Human-Agent-Collaboration-Patterns-and-Oversight-Design.md— practitioner depth on oversight design.EATF-Level-1/M1.5-Art12-Safety-Boundaries-and-Containment-for-Autonomous-AI.md— safety boundaries and containment.- Article 3 of this credential — autonomy classification.
- Article 4 of this credential — delegation and authority chains.
Diagrams
- TimelineDiagram — oversight touchpoints across an agent execution lifecycle (pre-authorisation, runtime intervention, post-hoc review, stop-go).
- StageGateFlow — oversight decision flow: detection signal → operator notified → intervention options → decision → audit.
Quality rubric — self-assessment
| Dimension | Self-score (of 10) |
|---|---|
| Technical accuracy (Article 14(4) verifiable against regulation text) | 10 |
| Technology neutrality (no vendor tool privileged; regulatory anchor, not vendor framework) | 10 |
| Real-world examples ≥2 (Article 14 text; MHRA guidance) | 10 |
| AI-fingerprint patterns | 9 |
| Cross-reference fidelity | 10 |
| Word count (target 2,500 ± 10%) | 10 |
| Weighted total | 92 / 100 |