01The Regulatory Landscape
The EU AI Act establishes a risk-based classification system with mandatory requirements for high-risk AI applications — including record-keeping, human oversight, and transparency obligations. The Colorado AI Act requires deployers of high-risk AI to implement risk management programs and provide notice when AI is used in consequential decisions. NIST AI RMF provides a voluntary framework that is rapidly becoming a de facto standard for organizations seeking structured AI governance.
Example: Regulatory overlap in financial services
A US bank operating in the EU uses AI for credit decisioning, customer support automation, and fraud detection. The credit decisioning AI falls under the EU AI Act's high-risk category, requiring transparency records, human oversight documentation, and conformity assessments. The same system is subject to the CFPB's adverse action notice requirements under US fair lending law. Fraud detection AI must comply with BSA/AML record-keeping requirements. Customer support AI falls under state-level consumer protection rules. Four distinct regulatory frameworks apply to three AI systems — and each framework demands evidence, not documentation.
These frameworks share a common assumption: that organizations deploying AI systems can produce evidence of compliance on demand. Not documentation describing intended compliance. Evidence demonstrating actual compliance — records of AI behavior, attribution of decisions, proof of human oversight.
02Documentation vs Infrastructure
There is a persistent gap between compliance documentation and compliance infrastructure. Most organizations have the former. Few have the latter.
Example: The fire safety analogy in practice
A large retailer's AI team maintains a 40-page responsible AI policy that states: "Customer-facing AI interactions are reviewed for quality on a weekly basis." An auditor asks for evidence of the review process for the past quarter. The team produces spreadsheets showing that a quality analyst reviewed 150 randomly selected interactions per week — roughly 2% of the 7,500 weekly AI interactions. The auditor notes that the remaining 98% have no evidence of review, oversight, or outcome verification. The policy describes intent. The 2% sample describes effort. Neither demonstrates operational compliance.
Compliance infrastructure produces evidence as a byproduct of normal operations. Each AI-assisted task generates a Trusted Work Unit containing the full evidence chain: what happened, who did what, whether the outcome met standards, and a cryptographic seal ensuring the record has not been altered. The compliance evidence is the work record. The compliance report is a query against existing data.
The distinction is analogous to fire safety. A fire escape plan is documentation. A fire escape is infrastructure. Regulators increasingly expect the latter.
03Building Evidence Systems
Compliance infrastructure captures evidence at the point of work, not the point of audit. This requires:
The key architectural decision is whether to build evidence capture into the operational workflow or bolt it on after the fact. The former produces complete records. The latter produces samples and reconstructions — which auditors increasingly reject as insufficient.
04Continuous Compliance
Traditional compliance operates on assessment cycles: annual audits, quarterly reviews, periodic attestations. AI compliance must operate continuously because AI systems operate continuously.
Example: Continuous vs periodic detection
A healthcare payer uses AI to pre-authorize medical procedures. Under periodic compliance, a quarterly review examines 500 pre-authorization decisions out of 45,000. The sample shows 97% accuracy. Between reviews, a model drift causes the AI to systematically deny physical therapy authorizations for patients with specific diagnosis codes — affecting 2,100 patients over eight weeks. Under continuous compliance, each pre-authorization generates a sealed TWU. A policy rule monitors denial rates by procedure type and diagnosis code. When physical therapy denials spike 340% for the affected diagnosis codes within 72 hours of the model update, the system flags the anomaly. The compliance team investigates within days, not months.
The only viable approach is infrastructure that produces audit-ready records as work happens. This shifts compliance from a periodic assessment to a continuous state. The organization is not "preparing for compliance." It is "operating in compliance" — continuously, verifiably, with evidence.
05Cross-Framework Compliance
Organizations operating across jurisdictions face overlapping regulatory requirements. The EU AI Act, Colorado AI Act, NIST AI RMF, ISO 42001, and sector-specific frameworks (HIPAA, SOX, PCI-DSS with AI extensions) all impose requirements that overlap but do not align perfectly.
Example: Single evidence, multiple frameworks
A multinational BPO processes customer interactions for clients in the US, EU, and UK. A single AI-assisted customer interaction in the UK must satisfy: EU AI Act transparency requirements (evidence of AI involvement), UK ICO data protection requirements (evidence of lawful processing), the client's SOC 2 controls (evidence of access controls and data handling), and the BPO's internal quality framework. Without unified evidence infrastructure, the BPO maintains four separate compliance tracking systems — each capturing partial, overlapping data. With TWU-based infrastructure, the sealed work record for each interaction contains the complete evidence chain. Compliance reports for each framework are different queries against the same underlying data.
Building separate compliance programs for each framework is operationally unsustainable. The alternative is infrastructure that produces granular evidence — sealed work records with attribution, outcome classification, and oversight documentation — that can be queried and formatted for any framework's specific requirements.
This is a core function of governance infrastructure: producing universal evidence that can be translated into framework-specific compliance reports. The evidence is the same. The reporting adapts. See also practical guidance on auditing AI agents within these frameworks.
