Skip to main content
AITF M1.8-Art13 v1.0 Reviewed 2026-04-06 Open Access
M1.8 M1.8
AITF · Foundations

Logging, Auditing, and SIEM Integration for AI Systems

Logging, Auditing, and SIEM Integration for AI Systems — AI Use Case Management — Foundation depth — COMPEL Body of Knowledge.

8 min read Article 13 of 15

This article walks the AI logging surface, the integration patterns with Security Information and Event Management (SIEM) platforms, and the audit requirements that distinguish AI logging from generic application logging.

The AI logging surface

A production AI system emits logs at six distinct points.

Inference logs record every inference request and response, with the authenticated caller, the model version, the input (or its hash where the input is sensitive), the output (or its hash), the latency, the policy decisions taken at the gateway (Article 6), and the validation outcomes (input out-of-distribution score, output content-filter result, schema-validation status). Inference logs are the primary forensic artefact for any incident affecting the production behaviour of the system and are the primary input to security detection and to model-quality monitoring.

Model lifecycle logs record every model load, every model artefact promotion across environments, every signature verification (Article 4 and Article 12), and every retirement. The lifecycle log answers the audit question “which model was running in production at the time of the incident?” with cryptographic-grade evidence.

Training logs record every training run with the source data version, the code version, the hyperparameters, the resource usage, the validation metrics, and the resulting model artefact. Training logs support the reproducibility audit, the regulatory documentation (the EU AI Act technical-documentation requirements explicitly contemplate this), and the supply-chain investigation when an upstream compromise is later discovered.

Access and identity logs record every authentication, authorization decision, secret retrieval (Article 7), and privileged access to AI infrastructure. The logs feed both the standard cybersecurity monitoring and the AI-specific monitoring for credential abuse and credential drift.

Network logs record connection patterns within and across the AI workload zones (Article 8). Mesh telemetry is the operational source; the SIEM is the analytic consumer. Network logs support detection of lateral movement, data exfiltration, and policy violations.

Operational logs record the platform-level events — pod restarts, autoscaling events, configuration changes, infrastructure-as-code applications. The operational log is shared with the broader platform’s logging rather than being AI-specific, but the AI workloads should be tagged so that AI-relevant operational events can be correlated with AI-specific logs during investigation.

The volume implication of comprehensive logging is significant. A production LLM system at moderate scale may emit terabytes of inference log per day, and the cost of retaining the volume forever is prohibitive. Mature programs implement tiered retention: full-fidelity hot retention for short windows (days to weeks), aggregated warm retention for longer windows (months), and sampled cold retention for the longest windows (years, primarily to support audit and regulatory inquiry). The retention tiers are themselves a governance decision driven by the regulatory framework, the audit posture, and the threat model from Article 1.

SIEM integration

The Security Information and Event Management platform — Splunk, Microsoft Sentinel, Elastic Security, Sumo Logic, the cloud-platform-native SIEMs, the next-generation security-data-lake platforms — is the analytic destination where the logs converge for security detection. AI logs integrate with the SIEM through the same patterns the rest of the security telemetry uses, with AI-specific extensions where the data warrants.

The integration pattern that works has three components.

Structured emission. AI logs are emitted in structured form (JSON, Protocol Buffers, OpenTelemetry events) with consistent field names and schemas across the AI estate. The structure enables the SIEM to index, correlate, and query the data efficiently. Free-text logs that require parsing in the SIEM are an anti-pattern that consumes capacity and reduces detection effectiveness.

Reliable shipping. Logs are shipped to the SIEM through a transport that survives partial failures of either the source or the SIEM. The pattern uses a buffered shipper (Fluent Bit, Vector, the cloud-native equivalents) at the source that persists logs through transient SIEM unavailability and replays them when connectivity restores. Lost logs are missed detections and missed audit evidence.

Detection content. The SIEM is configured with detection rules and analytics specific to the AI threat model — high-rate query patterns suggestive of model extraction (Article 4), input out-of-distribution score spikes suggestive of adversarial campaigns (Article 2), prompt-injection signature matches (Article 3), credential anomalies suggestive of theft (Article 7), and the broader catalog of AI-specific attack patterns from MITRE ATLAS https://atlas.mitre.org/. The detection content is itself a maintained artefact that evolves with the threat landscape.

The NIST AI Risk Management Framework Cybersecurity profile https://www.nist.gov/itl/ai-risk-management-framework prescribes logging and monitoring as a managed practice for AI systems. NIST SP 800-218A https://csrc.nist.gov/pubs/sp/800/218/a/final names structured logging and SIEM integration as required Secure Software Development Framework practices. The OWASP Top 10 for Large Language Model Applications https://owasp.org/www-project-top-10-for-large-language-model-applications/ catalogs Insufficient Logging and Monitoring as a contributing factor to several specific LLM vulnerabilities and the cure is the structured-emission, reliable-shipping, detection-content pattern above.

Audit requirements specific to AI

AI-specific regulations and management-system standards add audit requirements beyond what generic application logging satisfies.

The EU AI Act, Article 15 https://artificialintelligenceact.eu/article/15/ requires high-risk AI systems to log their operation in ways that support traceability. The implementation is system-specific — for some classes of system the requirement is a per-inference audit record, for others it is aggregate operational telemetry — but the regulator expects the operator to demonstrate that the logging design meets the traceability obligation. The Act’s technical-documentation provisions also require the operator to demonstrate that the logs are protected against tampering — an integrity requirement that drives the use of write-once or hash-chained log storage.

ISO/IEC 42001:2023 Annex A.7 https://www.iso.org/standard/81230.html requires AI Management System operators to establish logging and monitoring controls covering the AI system lifecycle, with explicit reference to inference activity, model updates, and access to AI assets. The certification audit reads the logging design and the retention policy and verifies that the controls operate as documented.

Sectoral regulations add requirements specific to the domain. Financial-services regulators expect inference activity for regulated decisions to be logged at the per-decision level for audit trail purposes. Healthcare regulators (the Health Insurance Portability and Accountability Act, the equivalent regulations in other jurisdictions) require access logging for systems that handle protected health information. The compliance mappings in Article 15 of this module spell out the specific requirements per framework.

The Gartner AI TRiSM Hype Cycle https://www.gartner.com/en/articles/gartner-top-strategic-technology-trends-for-2024 tracks the maturity of AI-specific observability and audit-evidence tooling, increasingly distinguishing the category from generic application observability as the AI-specific requirements harden.

Maturity Indicators

Foundational. AI workloads emit unstructured logs to local files or to a generic application logging stream. There is no SIEM integration. Inference activity is not retained at sufficient fidelity for audit. The team cannot reconstruct what happened on any specific past inference request. Audit evidence is produced reactively when a question is asked.

Applied. AI workloads emit structured logs into a centralized logging pipeline. Inference activity is retained at sufficient fidelity for short-term operational use. The SIEM ingests at least the highest-priority AI logs. Basic detection content exists for credential anomalies and obvious abuse patterns.

Advanced. Comprehensive AI logging covers the six log surfaces above with consistent structure across the estate. Tiered retention is implemented with policy that satisfies the regulatory framework. The SIEM ingests all AI logs and detection content addresses the AI-specific threat catalog. Log integrity is protected against tampering. The threat model from Article 1 names insufficient logging as a vulnerability and the controls map back to it.

Strategic. Logging and audit are first-class governance surfaces. Audit evidence is produced on demand from the logging platform. Detection content is curated as a maintained artefact and evolves with the threat landscape. Anomaly detection on AI-specific signals (extraction patterns, prompt-injection campaigns, credential drift) feeds incident response (Article 14) on a tracked schedule. The logging posture is itself audited on a regular schedule by external specialists.

Practical Application

A team operating AI workloads with insufficient logging should make three changes this quarter. First, audit which AI workloads emit logs to where, with what structure, and at what retention; the audit will surface workloads whose logging is inadequate for either operational use or audit. Second, deploy a structured logging pipeline (OpenTelemetry, Fluent Bit, or the cloud-native equivalent) for at least the highest-stakes production workloads, emitting inference logs at the fidelity the threat model and the regulatory framework require. Third, integrate the AI logging stream into the existing SIEM with at least a starter set of detection rules covering credential anomalies, query-rate anomalies, and input-validation failures.

These three actions create the logging foundation on which audit evidence, security detection, and incident response are subsequently built. They also produce the artefacts that the compliance mappings in Article 15 require for AI workloads under modern regulatory frameworks.


© FlowRidge.io — COMPEL AI Transformation Methodology. All rights reserved.