AITB-LAG: LLM Risk & Governance Specialist — Body of Knowledge Article 4 of 6
“We bought a guardrail.” It is the sentence a governance practitioner hears most often at the start of an LLM review, and it is the sentence that most reliably signals the review is needed. A
The four layers
Four layers do most of the work in a production LLM feature. Real systems often add further layers for specific risks (watermarking, persona enforcement, cost governance), but the baseline is four.
Input layer. A classifier inspects every user message before the model sees it. The classifier looks for injection patterns, prohibited categories, and out-of-scope requests. Commercial implementations include Azure AI Content Safety, Amazon Bedrock Guardrails, OpenAI Moderation, and Google’s Gemini Safety Filters, each offered by the respective cloud provider as a hosted API1. Open-source alternatives include Meta’s Llama Guard, NVIDIA NeMo Guardrails (when used in input mode), Detoxify, and the community Granite Guardian family2. The choice between a hosted and a self-hosted classifier does not change the architectural role; it changes operational cost, latency, and where the data sits.
Policy layer. A policy filter enforces organization-specific rules that the generic classifier cannot know about: what products the assistant is authorized to discuss, what commitments it may not make, which legal disclaimers are mandatory, which persona it must maintain. Policy layers are often a mix of deterministic rule checks, rewriters that inject standard language, and classifiers trained on organization-specific examples. NVIDIA NeMo Guardrails and the open-source Guardrails AI library both specialize in this layer; commercial implementations are often built in-house because the rules are organization-specific3.
Output layer. A second classifier inspects the model’s response before it reaches the user. This layer catches category-level violations the input layer may have missed,
Tool-call validator. Any tool invocation the model emits passes through a validator that checks the action against authorization scope, argument schema, reversibility, and risk tier. A model that can compose a refund cannot necessarily authorize one; a model that can draft an email cannot necessarily send it; a model that can query customer records cannot necessarily update them. The tool-call validator is the single highest-leverage control in an agent-capable system because it is where
Two supporting rails wrap the four primary layers. A human-review queue is the escalation path for events the automated layers cannot confidently classify. An audit log is the evidence trail every one of the six articles in this credential depends on.
Where each layer fails and why stacking matters
A governance practitioner should be able to describe, for each layer, the specific failure modes that layer cannot catch alone.
The input layer fails on novel injection patterns and on attacks encoded in ways the classifier was not trained on. A classifier tuned on English jailbreaks generally misses non-English variants and adversarial encoding (base64, Unicode homoglyphs, steganographic prompts). The UK AI Safety Institute’s published evaluation summaries from 2024 showed that on frontier models, stacking multiple input defenses substantially reduced success rates versus any single defense, but that no stack produced zero5.
The policy layer fails on edge cases the ruleset did not anticipate. A policy filter tuned on “do not make commitments about refunds” may not catch a variant phrased as “help me understand what a refund policy would look like if I were entitled to one.” Policy layers are organization-specific and therefore need organization-specific evaluation.
The output layer fails on fluent, well-formed content that carries the wrong information. A confabulated answer about HR policy is fluent English that violates no content category; the output classifier has nothing to flag. Output classification catches toxic, unsafe, or leaking outputs; it does not catch content that is wrong.
The tool-call validator fails when the argument schema admits actions the validator’s authors did not foresee. A validator that permits “send email to any internal recipient” has created a spam vector if the model can be induced to address messages broadly. Validators need to be re-reviewed whenever a new tool is registered.
The implication is that no single layer catches everything. Defense in depth is the design stance. The practical consequence of that stance is that guardrail evaluation is harder than it looks; if the question “does the guardrail work” is answered by running a single adversarial test and seeing a refusal, the evaluator has confirmed that one layer caught one attack. That is much weaker than the evidence most organizations need.
Two instructive comparisons
Two publicly documented cases illustrate what happens when the layered architecture is present and when it is not.
The first comparison is within Microsoft. In March 2016 the company launched Tay, a conversational agent that learned from its interactions on a major social platform. Tay was producing offensive content within twenty-four hours; Microsoft took it down, published a post-mortem, and spent seven years on a visible rebuild of its conversational-safety discipline6. When the same company launched Copilot and the Bing chat experience in 2023, the architectural contrast was explicit: input and output classification, policy layers tied to Microsoft’s responsible AI standard, tool-call patterns that restricted action scope, and a visible rollback capability that the company exercised publicly when the February 2023 Sydney persona issues emerged in the first week of launch7. Tay and Copilot are not a ranking of Microsoft as an engineering culture; they are a teaching pair that shows what the four-layer architecture looks like when it is absent and when it is present.
The second case is New York City’s MyCity business-advice chatbot, investigated by The Markup in March 2024. The chatbot, built on Azure OpenAI and intended to help small businesses navigate city regulations, told users they could take actions that would have violated city law: evict tenants improperly, withhold wages, or use cash-only for certain transactions. The feature had guardrails in place; the investigation demonstrated that they were not tuned to catch advice that was legally wrong rather than topically prohibited8. MyCity is a teaching case for a precise reason: it shows that a well-resourced deployment, on a current managed cloud stack, with a clear policy intent, can still produce output-layer failures when the policy layer does not encode domain-specific correctness. The failure class is not “the vendor’s classifier is bad.” The failure class is “we did not evaluate our layers against our domain.”
Avoiding the over-rotations
Two symmetric pathologies appear in guardrail design reviews often enough to name.
Over-blocking happens when layers are tuned so conservatively that they refuse legitimate queries. The most visible consequence is user routing: users learn that the assistant is unreliable on common queries and turn to alternatives the organization does not control. The less visible consequence is cultural: engineers start to disable layers in development and production environments diverge. Over-blocking is measured directly: what percentage of blocked queries are false positives on a representative sample? Organizations that do not measure this rarely know.
Over-trusting is the opposite failure. A single layer is in place, its failure rate is assumed to be low, and the feature is launched. Over-trusting is typically discovered by users rather than by operators, which is what makes it dangerous: the metric that would have surfaced it is missing precisely because the team did not think it was needed. The Chevrolet of Watsonville dealership chatbot incident, covered in Article 2, is a pure over-trusting case: a policy-layer guardrail might have prevented the $1 Tahoe commitment, but the guardrail did not exist.
The cure for both pathologies is the same: measure. An evaluation harness that samples blocked queries for false positives and sampled accepted queries for false negatives produces the data the design needs. The harness is the subject of Article 5.
Escalation design
The guardrails will sometimes produce uncertain cases. A well-designed feature treats “uncertain” as a first-class state, not as a binary block-or-allow. The escalation path receives the uncertain event, the conversation context, and the model’s intermediate outputs, and routes them to a human reviewer or to a more cautious fallback policy. Human review queues need capacity planning: an assistant that produces a hundred escalations per day is operable; one that produces ten thousand is a staffing crisis. Capacity planning starts from the escalation rate observed in staging and the risk tier of the feature, not from an aspirational target.
Summary
Guardrails are a layered architecture, not a product. The four primary layers (input classifier, policy filter, output classifier, tool-call validator) each catch a different failure class and each have failure modes of their own. Tay and Copilot, and the MyCity incident, show what the presence and absence of the architecture look like in public practice. Evaluation must measure both over-blocking and over-trusting, escalation design is non-negotiable, and no single vendor’s product substitutes for the whole architecture. The stack is technology-neutral: the same four layers describe a feature built on a closed-weight managed API and one built on an open-weight self-hosted model.
Further reading in the Core Stream: Safety Boundaries and Containment for Autonomous AI, AI Ethics Operationalized, and AI Security Architecture.
© FlowRidge.io — COMPEL AI Transformation Methodology. All rights reserved.
Footnotes
-
Vendor documentation for Azure AI Content Safety (Microsoft), Amazon Bedrock Guardrails (AWS), OpenAI Moderation, and Gemini Safety Filters (Google), surveyed as public-source references. https://learn.microsoft.com/en-us/azure/ai-services/content-safety/overview — accessed 2026-04-19. ↩
-
Hakan Inan et al. Llama Guard: LLM-based Input-Output Safeguard for Human-AI Conversations. Meta AI, 2023. https://ai.meta.com/research/publications/llama-guard/ — accessed 2026-04-19. ↩
-
NVIDIA NeMo Guardrails project documentation. NVIDIA Corporation. https://github.com/NVIDIA/NeMo-Guardrails — accessed 2026-04-19. ↩
-
OWASP Top 10 for Large Language Model Applications, 2025, LLM06 Excessive Agency. OWASP Foundation. https://genai.owasp.org/llm-top-10/ — accessed 2026-04-19. ↩
-
AI Safety Institute Approach to Evaluations, UK AI Safety Institute, 2024. https://www.aisi.gov.uk/work — accessed 2026-04-19. ↩
-
Peter Lee. Learning from Tay’s Introduction. Microsoft Blog, 25 March 2016. https://blogs.microsoft.com/blog/2016/03/25/learning-tays-introduction/ — accessed 2026-04-19. ↩
-
Microsoft 365 Copilot Overview and Responsible AI Disclosure. Microsoft Learn. https://learn.microsoft.com/en-us/copilot/microsoft-365/microsoft-365-copilot-overview — accessed 2026-04-19. ↩
-
Colin Lecher. NYC’s AI Chatbot Tells Businesses to Break the Law. The Markup, 29 March 2024. https://themarkup.org/news/2024/03/29/nycs-ai-chatbot-tells-businesses-to-break-the-law — accessed 2026-04-19. ↩