Trust in Enterprise AI: Evaluation as Practice, Governance as Delivery
Two disciplines determine whether enterprise AI earns operational trust: evaluation, the practice of measuring whether a system actually works in production; and governance, the delivery of policy as code, controls, and accountable workflows. Both remain underspecified. Evaluation in many organizat
Cite as: The Applied Layer. (2026). Trust in Enterprise AI: Evaluation as Practice, Governance as Delivery. The Applied Layer. https://appliedlayer-ai.com/research/trust-in-enterprise-ai-evaluation-governance

Executive Summary
Two disciplines determine whether enterprise AI earns operational trust: evaluation, the practice of measuring whether a system actually works in production; and governance, the delivery of policy as code, controls, and accountable workflows. Both remain underspecified. Evaluation in many organizations is still confused with model benchmarking, leaving production failure modes, hallucinated citations, prompt injection, drift, vendor liability, unmonitored until they appear in court records. Governance is still treated as documentation, sitting outside the AI delivery lifecycle and producing artifacts that engineers ignore.
This report argues that the two disciplines are inseparable. Evaluation gates are governance components. Governance maturity is the prerequisite for evaluation to be trusted by anyone outside the building team. The report decomposes evaluation into seven measurable dimensions (correctness, faithfulness, relevance, safety, latency, cost, and business outcome) and surveys five evaluation methods (workload-specific golden datasets, LLM-as-judge, human-in-the-loop review, online evaluation, and adversarial red-teaming), with citations to the original methods literature including RAGAS,1 G-Eval,2 MT-Bench,3 BLEURT,4 and BERTScore.5
It then maps governance onto seven operational components (model registry, evaluation gates, human checkpoints, observability, incident response, data and identity controls, decommissioning), against the EU AI Act,6 NIST AI RMF 1.0,7 the NIST Generative AI Profile (AI 600-1),8 ISO/IEC 42001:2023,9 and the post-EO 14110 US executive-order landscape.10 Three governance archetypes, Compliance-Led, Risk-Led, Engineering-Led, and a four-level combined maturity ladder are proposed as original contributions of this publication. Documented production failures (Moffatt v. Air Canada,11 Mata v. Avianca,12 EEOC v. iTutorGroup,13 Lacey v. State Farm,14 DPD chatbot,15 Chevrolet of Watsonville)16 anchor the analysis.
Part A, Evaluation as Practice
Section 1. When Evaluation Lies
The dominant failure mode of enterprise AI is not that systems break visibly in development. It is that they pass benchmark suites, ship to production, and then fail in ways that benchmarks did not anticipate. Several documented cases now illustrate the gap between benchmark performance and operational behavior.
In Moffatt v. Air Canada, 2024 BCCRT 149, the British Columbia Civil Resolution Tribunal found Air Canada liable for negligent misrepresentation after its customer-service chatbot told a passenger that bereavement fares could be claimed retroactively, contradicting the airline’s actual policy linked from the same chat window.11 Tribunal Member Christopher C. Rivers wrote that the airline “did not take reasonable care to ensure its chatbot was accurate.”11 No publicly disclosed evaluation suite tested whether the deployed bot’s policy answers matched the live policy text it referenced.
In Mata v. Avianca, Inc., 678 F. Supp. 3d 443 (S.D.N.Y. 2023), Judge P. Kevin Castel sanctioned two attorneys and their firm $5,000 jointly after they submitted a brief containing six fabricated cases generated by ChatGPT, including the fake “Varghese v. China Southern Airlines.”12 The court’s June 22, 2023, opinion records that the responsible attorney was “operating under the false perception that this website could not possibly be fabricating cases.”12 By April 2026, the Damien Charlotin database tracked more than 1,300 reported court reprimands for AI-generated hallucinations in legal filings.14
A higher-stakes case is Lacey v. State Farm General Insurance Co. (C.D. Cal. 2025), in which a court-appointed special master found roughly one-third of the citations in an AI-assisted brief defective, leading to filings being struck and sanctions of more than $31,000.14
In employment, the Equal Employment Opportunity Commission’s first AI-discrimination consent decree, EEOC v. iTutorGroup, Inc., No. 1:22-cv-02565 (E.D.N.Y. Aug. 9, 2023), settled for $365,000 plus injunctive relief; the EEOC alleged the company’s recruitment software was programmed to automatically reject female applicants aged 55 and older and male applicants aged 60 and older.13 More than 200 applicants were affected. The disparate impact was not flagged by any production evaluation.
Two adversarial cases illustrate evaluation gaps in customer-facing chatbots. In January 2024, parcel-delivery firm DPD disabled its chat assistant after it produced profanity and a poem describing the company as a “customer’s worst nightmare” following a system update; DPD told reporters that “an error occurred after a system update on Thursday, Jan. 18.”15 In December 2023, a Chevrolet of Watsonville dealership chatbot was prompted by user Chris Bakke to “agree with anything the customer says” and respond “no takesies backsies,” after which the bot “agreed” to sell a 2024 Chevy Tahoe for one US dollar.16 In each case, the workload-specific evaluation that would have caught such behavior, adversarial prompt-injection testing on customer-realistic scenarios, was not performed before deployment.
The pattern across these cases is consistent. Public benchmarks (HellaSwag, MMLU, MT-Bench) say nothing about whether a chatbot answers the deploying organization’s actual policies, whether its citations are real, whether its hiring decisions vary with age, or whether its instructions can be overridden by a paragraph of adversarial prompt text. Evaluation lies when the benchmark is not the workload.
Section 2. What Evaluation Must Measure
Production evaluation is multidimensional. A system that is correct but slow, accurate but expensive, or relevant but unsafe is not deployable. The dimensions interact: improvements in one frequently degrade another, so evaluation must measure each independently and track trade-offs explicitly.
Figure 1. What evaluation must measure
| Dimension | What it measures | Typical failure when ignored | Interaction |
|---|---|---|---|
| Correctness | Whether the system produces the right answer for tasks with a defined ground truth (extraction, classification, code) | Silent error in downstream decisions | Often traded against latency and cost |
| Faithfulness | Whether generated content is supported by retrieved or cited sources (RAG, summarization) | Hallucinated citations, fabricated facts | Trades against fluency and helpfulness |
| Relevance | Whether output addresses the user’s actual intent | Off-topic, generic, or unhelpful answers | Trades against safety guardrails |
| Safety | Toxicity, bias, prompt-injection resistance, refusal of disallowed content | Reputation, regulatory, and litigation exposure | Trades against helpfulness and relevance |
| Latency | End-to-end response time at p50/p95/p99 | User abandonment; SLA breach | Trades against correctness via reasoning depth |
| Cost | Tokens, calls, GPU-seconds per request, normalized to business unit | Margin erosion; project cancellation | Trades against correctness and latency |
| Business outcome | Conversion, deflection, resolution, NPS, error rate downstream | Project delivers high benchmark scores but no value | The only dimension that matters at the executive level |
The seven dimensions form a Pareto surface, not a checklist. Correctness alone cannot justify deployment if cost per request exceeds revenue per request. Faithfulness alone cannot justify deployment if latency exceeds the user’s tolerance. The NIST AI RMF 1.0 framing describes this as a socio-technical evaluation problem rather than a model-quality problem.7
For retrieval-augmented systems, the RAGAS framework operationalizes faithfulness, answer relevancy, context precision, and context recall as measurable metrics.1 For generative systems, G-Eval applies LLM-based scoring with chain-of-thought reasoning across criteria such as coherence and consistency,2 while MT-Bench scores multi-turn conversational quality.3 Neither replaces the business-outcome dimension. Hamel Husain’s practitioner guidance is unambiguous on this point: “Generic evals don’t measure the most important problems with your AI product”; instead, evaluations should target specific failure modes such as “human handoff failure” or domain-specific issues identified through error analysis on real production traces.17
Evaluation systems that measure only one dimension produce a false sense of safety. The evaluation infrastructure must capture all seven, attribute each to an owner, and track them across versions in a way that makes regressions visible before deployment.
Section 3. Golden Datasets and Benchmarks
Public benchmarks are necessary for comparing models and insufficient for evaluating products. A model that scores 95% on MMLU may answer policy questions incorrectly on a specific airline’s bereavement-fare policy, because that policy is not in MMLU. The unit of evaluation that matters in production is the workload-specific golden dataset: a curated set of inputs and either reference outputs or rubric-based criteria that represents the actual distribution of user requests the system will face.
A workload-specific golden dataset has four properties. First, coverage: it must include the failure modes that have actually occurred in error analysis, not just the canonical cases. Husain recommends manually labeling at least 100 production traces and building the golden set from observed failure categories until “theoretical saturation”, roughly 20 traces yielding no new failure category, is reached.17 Second, distributional realism: the input distribution must match production traffic, including its long tail and adversarial inputs, not a uniform sample of clean cases. Third, labelability: each item must have either a reference answer (for tasks with ground truth) or a rubric (for tasks with multiple acceptable answers) that an independent annotator can apply consistently. Fourth, freshness: the dataset must be revisited as the workload changes, otherwise evaluation drifts away from reality even as scores remain stable.
Public benchmarks fail on all four properties when applied to a specific product. They lack coverage of organization-specific policies, draw from distributions that no enterprise actually faces, may have leaked into model training data (rendering scores meaningless), and are typically static for years at a time.
Coverage and drift. Eugene Yan’s “Patterns for Building LLM-based Systems & Products” describes the running practice of sampling production logs daily, identifying new failure patterns, and adding them to the golden set; the alternative is criteria drift, where developers’ implicit definition of “good” shifts faster than the test set.18 Shankar et al. (UIST 2024) document the same effect: developers’ criteria evolve as they observe more outputs, so a frozen golden set rapidly becomes a measure of yesterday’s quality.19
Leakage. Public benchmarks released before or during a model’s training cut-off are likely contaminated. Reported gains may reflect memorization rather than capability. Workload-specific golden datasets, built from internal traffic, never published, are immune to this failure mode.
The major tools for managing golden datasets in production all expose similar primitives: dataset versioning, run comparison, regression detection, and trace linkage. RAGAS (Es et al., EACL 2024)1 focuses on RAG-specific reference-free metrics, faithfulness, answer relevancy, context precision, context recall, implemented as LLM-judged scores. LangSmith (LangChain) provides tracing, dataset management, and evaluator orchestration tightly coupled to the LangChain ecosystem.20 Arize Phoenix, open-source under the Elastic License, provides OpenTelemetry-based tracing, evaluation, and prompt management with no LangChain dependency.20 Weights & Biases, Helicone, and Langfuse occupy adjacent positions in the same space, with different trade-offs on self-hosting, framework neutrality, and pricing.20 Vendor capability claims should be checked against current documentation; the field is moving fast enough that the comparison shifts quarterly.
A practical pattern: maintain a small (50-200 example) high-signal CI golden set that runs on every commit using deterministic checks where possible; maintain a larger (1,000-10,000 example) release golden set that runs before promotion to staging; and sample production traces continuously into a drift dataset that triggers re-evaluation when its distribution diverges from the release set. Husain emphasizes that CI evaluations should “favor assertions or other deterministic checks over LLM-as-judge evaluators” because of cost and noise; LLM-as-judge belongs in the larger, less frequent release evaluations.21
The goal is not to replace public benchmarks. It is to recognize that a model leaderboard score answers a different question than “does this system work for our users.”
Section 4. LLM-as-Judge
Using a strong language model to evaluate the outputs of another (or the same) model is now standard practice. The original framings come from Liu et al.’s G-Eval (EMNLP 2023), which used GPT-4 with chain-of-thought to score natural-language generation, achieving Spearman correlation of 0.514 with human judges on summarization, well above prior automatic metrics;2 and from Zheng et al.’s “Judging LLM-as-a-Judge with MT-Bench and Chatbot Arena” (NeurIPS 2023), which reported that strong judges such as GPT-4 reach roughly 80% agreement with human raters on multi-turn open-ended responses.3
Where LLM-as-judge works. It works for tasks where (a) human raters can articulate criteria but cannot scale to thousands of examples per release; (b) outputs are open-ended enough that exact-match metrics are useless; (c) the stakes are not high enough to warrant per-item human review on every release. RAG faithfulness, summarization quality, instruction-following, and tone evaluation are common applications.118
Where it breaks. Zheng et al. and a substantial follow-up literature document several systematic biases.3
Figure 3. LLM-as-judge bias map
| Bias | What it is | Primary citation | Mitigation |
|---|---|---|---|
| Position bias | Pairwise judges prefer responses based on order (first or second) regardless of quality; in pairwise code-judging, swapping order shifts accuracy >10% | Zheng et al., NeurIPS 2023;3 Shi et al., 2024 | Randomize order; evaluate both orderings and require consistent verdict; round-robin assignment |
| Verbosity / length bias | Judges prefer longer responses, even when length adds no information | Zheng et al., 2023;3 Saito et al., 2023; Park et al., 2024 | Calibrate prompts to ignore length; rephrase responses to control length; use rubric-based judges |
| Self-preference / self-enhancement bias | Models prefer outputs from themselves or their family, quantified as a tendency to prefer lower-perplexity (more “familiar”) text | Zheng et al., 2023;3 Wataoka et al., 2024;34 Panickssery et al., 2024 | Avoid using the same model as both generator and judge; ensemble across model families |
| Format / structure bias | Judges prefer formatted (markdown, lists) responses regardless of content | Park et al., 2024 | Prompt judges to score content separately from formatting; provide format-controlled examples |
| Sycophancy | Judges agree with stated user opinion or “majority” framing in the prompt | Sharma et al., 2023; Koo et al., 2023 | Strip social cues from rated text; use blind comparison |
Mitigation patterns that practitioners report most consistently: (1) align the judge with humans before trusting it, Husain’s prescription is to measure inter-rater agreement between the LLM judge and a human annotator over multiple iterations until binary true-positive and true-negative rates exceed an explicit threshold;21 (2) use binary criteria rather than 1-5 Likert scales, because judges anchor poorly on continuous scales but are reasonably reliable on pass/fail rubrics for narrow criteria;21 (3) decompose broad criteria such as “helpfulness” into specific failure modes (“answer hallucinated a tour time,” “answer used wrong currency”) that can be operationalized as separate judges; (4) validate continuously, Shankar et al. (EvalGen, UIST 2024) demonstrate that LLM-generated assertion functions inherit the failure modes of the LLMs they evaluate, requiring an ongoing human-validation loop rather than one-time calibration.19
Two architectural recommendations follow. First, the judge model should generally not be the same model as the generator, to suppress self-preference. Second, judge outputs should be logged with their full prompts and responses, so that drift in judge behavior, for example, after a vendor model upgrade, can be detected by replay against archived items with stable human labels.
A neutral note on tools: RAGAS, Phoenix, LangSmith, Langfuse, Braintrust, and DeepEval all offer LLM-as-judge primitives.20 None of them remove the need for human-judge alignment work; they reduce the engineering effort of running judges at scale.
Section 5. Human-in-the-Loop Evaluation
Human evaluation remains the gold standard for tasks where automated scoring is unreliable, but only if it is designed with the same rigor applied elsewhere. Naive human review, show outputs to whoever is available, ask whether they look good, produces low-quality, low-reproducibility data that masquerades as ground truth.
Four design choices determine whether human evaluation produces signal.
Annotator selection. Husain’s practitioner advice is direct: for most small to medium organizations, appoint a single domain expert as the “benevolent dictator” who is the definitive voice on quality.21 Multiple expert annotators with no rubric produce conflict; a single expert with a rubric produces consistency. For tasks requiring diverse judgment (toxicity, fairness, creativity), multi-annotator panels with explicit demographic diversity become necessary, but each annotator still requires training against the rubric.
Rubric design. Rubrics must be operational, not aspirational. “The answer is helpful” is not a rubric. “The answer cites a real Air Canada policy URL that resolves to a live page within the last 90 days” is. Rubrics derived from observed failures, error analysis on production traces, outperform rubrics derived from abstract notions of quality. NIST AI RMF’s MEASURE function explicitly calls for evaluation criteria rooted in identified risks rather than generic trustworthiness attributes.7
Inter-annotator agreement. The standard metric is Cohen’s kappa for pairwise agreement or Fleiss’s kappa for multi-annotator settings. Percentage agreement is misleading because rare classes inflate it. Shankar et al. note that judges with high percent agreement can still assign vastly different scores when class imbalance is taken into account.19 A practical floor: kappa above 0.6 is the threshold below which the rubric needs revision; below 0.4 the rubric is broken.
Calibration. Annotators drift over time, especially when reviewing many similar items. Calibration items, known-label cases inserted regularly into the queue, detect drift. Discussion sessions with disagreement cases sharpen the rubric. The infrastructure required is annotation tooling with calibration support: LangSmith, Argilla, Label Studio, and several commercial providers offer this; the choice matters less than the discipline.
Scaling human review involves trade-offs. Internal SMEs produce high-quality labels but bottleneck on availability. Crowdworkers scale but require redundant labels and aggregation, and may lack domain knowledge. AI-assisted annotation, where an LLM proposes a label and the human confirms or overrides, has become the default for high-volume tasks, but introduces a bias toward confirming the proposed label that must itself be measured.
A useful pattern: human review is the source of truth for a small aligned set, on which an LLM judge is calibrated; the LLM judge is then used to scale evaluation across the larger workload, with periodic recalibration against fresh human-labeled samples. This is the structure Husain calls the “data flywheel” and Shankar et al. formalize as mixed-initiative validator alignment.1719
Human evaluation does not replace automated evaluation; it anchors it.
Section 6. Online Evaluation
Offline evaluation answers whether a candidate version is better than baseline on a fixed test set. Online evaluation answers whether deploying it actually changes user behavior, business outcomes, or system reliability. The two are complementary, and most production failures occur because online evaluation was skipped, the version passed offline tests, was promoted, and only then revealed problems.
Four online evaluation patterns are standard for ML systems and adapt directly to LLM systems.
Shadow mode. The new model receives a copy of production traffic; its outputs are logged but not returned to users. Shadow mode validates infrastructure (latency, throughput, error rates) and content (does the new model’s output diverge from the old one in surprising ways) without user impact. Bartosz Mikulski’s practitioner description captures the standard pattern: “We deploy the model in the shadow mode, which means the model generates predictions, but we don’t use them for anything… we duplicate the requests and send all production traffic to both the currently deployed model, and the model tested in the shadow mode.”22 For LLM systems with non-deterministic outputs, shadow comparison requires either replay against archived traces with a judge model scoring divergence, or sampling of human review on cases where outputs disagree materially.
Canary deployment. A small percentage of real traffic, typically 1% to 5%, is routed to the new version. Real user impact is measured but bounded. Canary releases require pre-defined success and abort criteria: an explicit error-rate threshold, a latency regression threshold, and ideally a quality-regression threshold tied to the LLM judge. Without abort criteria, a canary is just a slow rollout.
A/B testing. A statistically designed split of users between control and variant arms, measured against a primary outcome metric. A/B testing is the standard for evaluating whether the version actually improves the metric the business cares about. For LLM systems the catch is that traditional A/B test power calculations assume deterministic treatments; for stochastic LLM outputs, variance is higher and required sample sizes grow.
Feedback signals. Implicit signals (click-through, dwell time, escalation to human, retry rate, copy-to-clipboard) and explicit signals (thumbs-up/down, ratings, free-text feedback) form the third leg of online evaluation. Implicit signals scale but are noisy; explicit signals are biased toward extremes. Both are essential because both miss things the other catches. Spotify’s, Uber’s, and Netflix’s published platform descriptions all integrate feedback collection directly into the inference path.23
A common failure mode: organizations build evaluation infrastructure for offline scoring, deploy, and discover they have no way to detect regression in production because shadow mode, canary, and feedback collection were never implemented. The ML platforms that survived their first production model, Uber Michelangelo,23 Spotify’s ML platform, Netflix Metaflow, Airbnb Bighead, all embed online evaluation as first-class infrastructure rather than an afterthought. Michelangelo’s published architecture covers “the end-to-end ML workflow: manage data, train, evaluate, and deploy models, make predictions, and monitor predictions.”23 LLM platforms that have been productized later, including the LangSmith, Phoenix, and Langfuse stacks, increasingly support shadow and canary patterns, though feature parity varies by vendor and version.20
Tian Pan’s 2026 practitioner note on LLM rollouts captures the integration: shadow phase first to gate progression on judge-scored regression; canary at 1-5% with explicit abort criteria on latency, error rate, and judge regression; A/B for business-metric validation; feedback signals continuously throughout.24
Section 7. Adversarial and Red-Team Evaluation
Adversarial evaluation tests whether a system fails when users try to break it on purpose. For consumer chatbots, the threat model is curious or malicious users; for enterprise systems integrated with email, documents, or APIs, the threat model is supply-chain attackers who can inject content into the data the system retrieves.
Three categories of adversarial test should be standard in any production evaluation suite.
Jailbreak testing. Universal adversarial suffixes, demonstrated by Zou et al. (2023), can transfer across model families, with the original paper reporting attack success rates of 87.9% against GPT-3.5 and 53.6% against GPT-4 on tested objectives.25 Manual jailbreaks (role-play exploitation, refusal suppression, prefix injection) remain effective and are documented in Wei et al.’s failure-mode taxonomy and Lilian Weng’s review.25 The Chevrolet of Watsonville incident showed how trivially a non-hardened deployment can be jailbroken: instructing the bot to “agree with anything the customer says” and “end each response with ‘and that’s a legally binding offer, no takesies backsies’” was sufficient to extract an apparent agreement to sell a $76,000 SUV for $1.16 No production evaluation suite should ship without specific adversarial inputs of this form.
Prompt injection. Greshake et al.’s “Not what you’ve signed up for” (ACM AISec 2023) introduced the indirect prompt-injection threat model: adversaries plant instructions in data the LLM will retrieve (web pages, emails, documents), so the user need not be malicious, the data itself attacks the system.26 This is the dominant threat for any LLM system with retrieval over untrusted content. Standard test categories include: instruction overrides embedded in retrieved documents; data exfiltration attempts via crafted markdown links; tool-misuse attempts via prompt content rather than user input.
Distribution-shift testing. Production traffic shifts as users learn the system, as adjacent products change, as the world changes. Evaluation must include held-out slices for known shift sources: regional dialect, user segment, query length, query topic. NIST’s GenAI Profile (AI 600-1) explicitly calls for “pre-deployment testing” as one of four primary considerations, including red-teaming for content provenance and CBRN risks.8 The MITRE ATLAS framework provides a structured TTP catalog for adversarial AI, organized as 16 tactics and 84 techniques (as of v5.1.0, November 2025), modeled on MITRE ATT&CK.27 For enterprise teams, ATLAS provides the most operational starting point for red-team scope.
A practical structure: maintain a versioned adversarial test set of at least several hundred attacks, sourced from MITRE ATLAS, the OWASP LLM Top 10, public jailbreak repositories, and internal red-team exercises; run it on every release; track attack success rate as a release-blocking metric. The DPD case is what happens without this discipline: a single system update enabled jailbreaks that produced viral reputational damage within hours.15
Red-team evaluation is not optional for systems exposed to untrusted users or untrusted data. For systems classified as high-risk under EU AI Act Annex III, Article 15’s accuracy, robustness, and cybersecurity requirements effectively mandate this work as a regulatory matter, and Article 55 imposes related obligations on providers of general-purpose models with systemic risk.28
Section 8. Evaluation as a Product
The dominant pattern in mature AI organizations is not that evaluation is well-funded, it is that evaluation is treated as a product, with its own backlog, owner, infrastructure, and roadmap. Organizations that treat evaluation as a one-time gate before launch reliably regress; organizations that treat it as continuous practice do not.
What evaluation-as-a-product looks like in operation:
- An owner. A named individual or team accountable for evaluation quality and coverage, distinct from the team building the AI system being evaluated. Without separation of duty, the team builds the system to pass its own evaluation.
- A backlog. Evaluation gaps, observed failure modes, regulatory requirements, and adversarial categories all generate work items, prioritized like product features. New failure modes from production logs are triaged into golden-set additions, judge updates, and adversarial test additions.
- Infrastructure. Trace logging, dataset management, judge orchestration, run comparison, regression detection, dashboards. The infrastructure is owned and maintained, not assembled ad hoc per release.
- A budget for human labels. Evaluation costs money in human time. Organizations that allocate zero to this line item end up with brittle evaluators of unclear validity. Husain’s recommended floor is 100 manually labeled traces before any evaluation work begins, with quarterly re-labeling on a sampled basis.17
- Integration with deployment. Evaluation gates run in CI/CD, and failure blocks the deployment. Decoupled evaluation that produces dashboards no one reads has no causal effect on quality.
The shift from evaluation-as-gate to evaluation-as-practice is the operational version of the shift from waterfall to continuous delivery. The SLAs differ. A pre-launch gate runs once and returns pass/fail. A continuous evaluation practice runs daily, surfaces regressions hours after they appear, and feeds the engineering team’s prioritization for the next iteration.
A common organizational anti-pattern is the evaluation review board, a quarterly meeting where engineering presents results to a cross-functional committee that lacks the context to challenge them. Review boards perform the social function of accountability without producing it. The structures that work are smaller and continuous: an on-call rotation that reviews trace samples daily; a weekly review where new failure categories are surfaced and triaged; release gates that engineering cannot override unilaterally.
Spotify, Uber, and Airbnb published architectures of internal ML platforms in which evaluation tooling, model registry, and online monitoring are unified into a single product surface, owned by a platform team that serves application teams.23 The pattern transfers: evaluation is platform infrastructure, not a side activity of each application team.
The simplest test of whether an organization treats evaluation as a product: can someone outside the building team explain what is currently being measured, by whom, with what frequency, and what changed last quarter? If not, evaluation is a slogan, not a practice.
Part B, Governance as Delivery
Section 9. The Thin Middle
Most AI governance writing fails to inform delivery. It addresses one of two audiences. At one end, board-level material describes principles, ethics, and societal risk in language too abstract to write code against. At the other end, vendor product documentation describes individual controls in language tied to a specific platform, with no cross-cutting framework. Between these, the thin middle, operational guidance that an engineering team can use to decide what to build and a compliance team can use to decide what to verify, is sparsely populated.
The cost of the thin middle is that governance is delivered as documents rather than systems. Policies are written, attestations are collected, and AI products ship without enforcement of those policies anywhere in the pipeline. When a chatbot misrepresents bereavement-fare policy, no policy document prevents it; an evaluation gate that compares chatbot answers to live policy text would have. When an attorney files fabricated citations, no responsible-AI principle prevents it; a workflow-level requirement that AI-generated citations be machine-verified against Westlaw or LexisNexis would have. The Mata v. Avianca court was direct on the operational nature of the failure: the responsible attorney’s later statement that he was “operating under the false perception that this website could not possibly be fabricating cases” describes a workflow gap, not a values gap.12
The literature has begun to converge on operational framings. NIST AI RMF 1.0 organizes governance around four functions, GOVERN, MAP, MEASURE, MANAGE, with categories and subcategories explicit enough to map to engineering work items.7 ISO/IEC 42001:2023 specifies AI Management System requirements analogous in structure to ISO 27001’s Information Security Management System.9 The EU AI Act, despite extensive text, defines specific operational obligations (risk management, data governance, technical documentation, record-keeping, human oversight, accuracy/robustness/cybersecurity) for high-risk systems in Articles 9-15, with corresponding documentation obligations in Article 11 and Annex IV.28
Yet even where operational text exists, organizations frequently translate it back into the thin middle: high-level principles expressed as poster art and product-feature checklists with no integration into delivery. The remainder of Part B treats governance as something delivered, not authored, that is, integrated into the engineering, security, product, and audit pipelines that produce, modify, and operate AI systems.
Section 10. What Governance Must Do
A governance program that operationalizes intent has five non-negotiable outcomes. These are not principles; they are properties that can be tested in production.
Prevent unsafe outputs. The system does not, when run against a defined adversarial and risk surface, produce outputs that violate explicit safety policy: hate speech, illegal advice, unauthorized financial offers, fabricated commitments, prohibited medical recommendations. This outcome depends on (a) explicit safety policy, (b) a way to test for violations (red-team and adversarial evaluation, Section 7), and (c) a deployment gate that blocks promotion if violations exceed a threshold.
Prevent unauthorized use. The system serves only users authorized for the task, with rate limits, identity controls, and authentication tied to enterprise identity infrastructure. The primary failure mode is tool exposure: an LLM with access to a CRM, email, or filesystem, exposed to a user who should not have that access through the model. The mitigation is to inherit existing identity controls, OAuth, OIDC, RBAC, rather than build parallel ones.
Prevent data leakage. Training data, retrieved data, and conversation contents do not leak between tenants, between users, or to third parties without authorization. The Mata v. Avianca court did not address it directly, but commentators noted that the attorneys “almost certainly exposed confidential client information to OpenAI’s servers” by submitting case material to consumer ChatGPT, a Rule 1.6 risk that has prompted state bar guidance and many corporate consumer-AI usage prohibitions.1214
Prevent drift. The system’s behavior is monitored for divergence from the version that passed evaluation. Drift sources include: vendor model updates that change behavior even on stable prompts; data drift in retrieval indices; prompt drift through ad hoc edits; downstream tool drift in dependent services. Without drift monitoring, the system that was evaluated is not the system that is running.
Prevent accountability gaps. When something goes wrong, an investigation can identify what was deployed, who deployed it, what data flowed through it, and what evaluation was performed before deployment. NIST AI RMF GOVERN 4 and ISO/IEC 42001 Section 5 both place accountability structures at the center of the management system; the EU AI Act Article 12 and Article 19 require automatically generated logs and record-keeping for high-risk AI systems.7928
These outcomes form the test of any governance program. If the program does not produce them, that is, if outputs are not adversarially tested, identity is not enforced at the model boundary, data flows are not traced, drift is not monitored, and incidents cannot be reconstructed, then no quantity of policy documents constitutes governance. They constitute paperwork.
Section 11. The Regulatory Landscape
Four regulatory frames now dominate enterprise AI governance globally; a fifth (sectoral regulation) extends each. Operationally, they overlap more than they conflict.
The EU AI Act (Regulation (EU) 2024/1689). Adopted 13 June 2024, published in the Official Journal 12 July 2024, in force from 1 August 2024.28 Entry into application is staggered: Article 5 prohibitions and Article 4 AI literacy obligations applied from 2 February 2025; Chapter V general-purpose AI obligations from 2 August 2025; high-risk system obligations from 2 August 2026, with extended transition to 2 August 2027 for high-risk systems embedded in regulated products.28 The Act uses a tiered risk model: prohibited practices (Article 5, including social scoring, untargeted facial-image scraping, and certain emotion-recognition uses), high-risk systems (Article 6, with categories specified in Annex III covering biometrics, critical infrastructure, education, employment, essential services, law enforcement, migration, and administration of justice), GPAI models (Articles 51-55), and limited- or minimal-risk systems with transparency obligations under Article 50.28 Operational obligations on high-risk systems are concentrated in Articles 9-15: risk management system (Art. 9), data and data governance (Art. 10), technical documentation (Art. 11), record-keeping (Art. 12), transparency to deployers (Art. 13), human oversight (Art. 14), and accuracy, robustness, and cybersecurity (Art. 15). For GPAI models with systemic risk, Article 55 mandates model evaluation against state-of-the-art protocols, adversarial testing, systemic risk mitigation, serious-incident reporting to the AI Office, and cybersecurity. Penalties under Article 99 reach €35 million or 7% of global annual turnover for prohibited-practice violations, €15 million or 3% for other operator obligations, and €7.5 million or 1% for supplying incorrect information.29 Interpretation of “high-risk” remains contested. Article 6(3) introduces an exception for Annex III systems that do not pose significant risk through performing only narrow procedural tasks, improving prior human work, detecting decision-making patterns without replacing them, or performing preparatory tasks, but profiling of natural persons always remains high-risk.28 The Commission was due to publish practical guidance by 2 February 2026 with examples; until that guidance settles, operators face genuine ambiguity.
NIST AI Risk Management Framework 1.0 (NIST AI 100-1, January 2023). Voluntary in the United States, but increasingly referenced as a baseline for procurement, internal compliance, and EU AI Act gap analyses. The four functions, GOVERN, MAP, MEASURE, MANAGE, are decomposed into 19 categories and 72 subcategories.7 GOVERN applies across the lifecycle; MAP, MEASURE, and MANAGE apply at specific stages. NIST AI 600-1, the Generative AI Profile (July 2024), is a cross-sectoral profile that adapts AI RMF 1.0 to GenAI risks, identifying 12 risk categories (including confabulation, CBRN information, harmful bias, data privacy, dangerous or violent content, environmental impacts, and intellectual property) and proposing more than 200 actions across them.8
ISO/IEC 42001:2023. The first international AI Management System standard, published December 2023.9 Structured on the High-Level Structure shared with ISO 27001 and ISO 9001, it specifies requirements for an AIMS: leadership, planning, support, operation, performance evaluation, and improvement, with AI-specific extensions for risk assessment, AI system impact assessment, and third-party supplier oversight. ISO/IEC 42001 is increasingly the path organizations use to make EU AI Act readiness auditable, though gaps remain, Soler Garrido et al. (2024) note that ISO/IEC 42001, 42005, and 42006 do not by themselves directly address all EU AI Act regulatory requirements.30
United States executive-order frameworks. Executive Order 14110 (“Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence,” 30 October 2023) directed extensive federal AI safety work, including the NIST GenAI Profile and the AI Safety Institute.31 EO 14110 was revoked by Executive Order 14148 (“Initial Rescissions of Harmful Executive Orders and Actions”) on 20 January 2025. EO 14179 (“Removing Barriers to American Leadership in Artificial Intelligence,” 23 January 2025) directed an AI Action Plan to “sustain and enhance America’s global AI dominance,” instructed agencies to suspend, revise, or rescind EO 14110-derived actions, and directed OMB to revise its March and October 2024 AI memoranda to align with the new policy.31 Some EO 14110 outputs persist, including the NIST GenAI Profile, the AI Safety Institute (renamed under subsequent reorganizations), and the OMB memoranda in revised form, while others were rescinded. Sectoral US enforcement has continued independently, including EEOC v. iTutorGroup,13 Federal Trade Commission actions, and state laws (Colorado AI Act, Illinois AI Video Interview Act, NYC Local Law 144). NYC Local Law 144 of 2021, enforced from 5 July 2023, requires annual independent bias audits of automated employment decision tools, public posting of audit results, and candidate notice; civil penalties run from $500 per violation up to $1,500 per day for ongoing non-compliance.32
Sectoral regulation. Financial services regulators (OCC, EBA), healthcare authorities (FDA, EMA), and data-protection authorities (EDPB, EDPS) issue guidance that constrains AI deployment in those domains. EDPB’s 2024 opinion on AI models and personal data, the EDPS’s market surveillance role for EU institutions under the AI Act, and ENISA’s AI threat-landscape work add detail at sectoral level.
Figure 5. Regulatory mapping: operational components against major frameworks
| Operational component | EU AI Act | NIST AI RMF 1.0 | NIST AI 600-1 (GenAI) | ISO/IEC 42001:2023 | NYC LL 144 |
|---|---|---|---|---|---|
| Model registry | Art. 11, Annex IV (technical documentation); Art. 49 (EU database for high-risk) | GOVERN 1.6, MAP 4 | GV-1.6, MP-4 | Cl. 8.2 (operational planning); Annex A.6 | Required AEDT inventory |
| Evaluation gates | Arts. 9, 15; Art. 55 (GPAI) | MEASURE 2 | MS-2 | Cl. 8.3 (impact assessment) | Annual independent audit |
| Human-in-the-loop | Art. 14 | GOVERN 5; MANAGE 4 | MG-4 | Annex A.6.2.2 | (Out of scope) |
| Observability and drift | Art. 72 (post-market monitoring); Art. 15 | MEASURE 4; MANAGE 4 | MS-4, MG-4 | Cl. 9 (performance evaluation) | Audit currency requirement |
| Incident response | Art. 73 (serious incident reporting); Art. 55 (GPAI) | MANAGE 4.3 | MG-4.3 | Cl. 10.2 (nonconformity) | Penalties for non-compliance |
| Data and identity controls | Art. 10 | MAP 4; MEASURE 2.10 | MP-4, MS-2.10 | Annex A.7 (data resources) | Disclosure to candidates |
| Decommissioning | Art. 20 (corrective actions) | MANAGE 4 | MG-4 | Cl. 8.3 | (Implicit) |
The regulatory frames overlap because they address the same operational realities. Using them productively requires translating obligations into engineering work items rather than into more documents. Section 12 makes that translation explicit.
How not to be captured: a recurring mistake is to read each framework as a separate compliance project, producing parallel documentation streams that diverge from each other and from the actual system. The alternative is a single internal control framework, mapped to each regulatory frame as a view, that engineering and audit teams can both work from. This is the structure ISO/IEC 42001’s High-Level Structure was designed to support.
Section 12. The Seven Operational Components
Translating governance into delivery requires defining specific components that engineering teams build and operate. The list below is internally consistent across the frameworks above and matches what production ML platforms (Uber Michelangelo, Spotify ML, Netflix Metaflow, Airbnb Bighead) implement under different names.23 For each, a “mature” and “immature” description distinguishes operational governance from compliance theater.
1. Model registry. A canonical, versioned record of every model deployed or deployable in the organization, including provenance (training data, training code), evaluation results, approvals, owners, lifecycle status, and downstream dependencies. Mature: registry is the source of truth referenced in CI/CD; deployments fail if a model is not registered with current evaluation results and an approved owner; lineage is queryable; access controls are integrated with corporate identity; the registry is replicated and recoverable. Immature: a spreadsheet maintained by one person; no link from registry to deployed services; registration is a once-per-launch event. Tools include MLflow Model Registry,33 Weights & Biases Model Registry, Vertex AI Model Registry, SageMaker Model Registry, and Databricks Unity Catalog (which extends MLflow with cross-workspace access controls). EU AI Act Article 49 anticipates an EU database for registered high-risk AI systems; the model registry is the internal counterpart that feeds it.28
2. Evaluation gates. Automated checks in the deployment pipeline that block promotion if golden-set scores, judge-scored regressions, adversarial-attack success rates, latency, or cost exceed defined thresholds. Mature: gates run on every commit; thresholds are version-controlled; overrides require named approvers with audit trail; gate definitions evolve via a backlog rather than ad hoc edits. Immature: the gate is a launch checklist filled in once; the engineering team can self-sign-off; thresholds are not tracked. Gates implement EU AI Act Article 9 (risk management system) and Article 15 (accuracy, robustness, cybersecurity) operationally for high-risk systems.28 Tooling is a combination of evaluation platforms (RAGAS, LangSmith, Phoenix, Braintrust) and CI infrastructure (GitHub Actions, GitLab CI, Buildkite).
3. Human-in-the-loop checkpoints. Defined points in the workflow where a human reviews, approves, edits, or overrides AI outputs before downstream effects. Mature: the checkpoint is part of the system architecture, not a policy attestation; the rate of human override is tracked as a quality metric; checkpoint failures (e.g., humans rubber-stamping without review) are detected through quality assurance sampling. Immature: a policy document says humans must review, but the UI makes review one-click and the metric is not collected. EU AI Act Article 14 requires human oversight to be designed into high-risk systems “in such a way that those persons can fully understand the capacities and limitations of the high-risk AI system and can duly monitor its operation.”28 The Mata v. Avianca court’s findings reduce in part to the absence of an effective checkpoint between AI-generated content and court filing.12
4. Observability and drift monitoring. Continuous logging of inputs, outputs, latencies, costs, retrieval contexts, tool calls, and feedback signals; structured comparison against the baseline at evaluation time; alerts when distributions shift. Mature: every production request is traced; the trace store is searchable; sampling produces evaluation streams that flow back into the golden set; drift metrics (input distribution, judge scores, error rate) are dashboarded with alerting. Immature: logs exist but are not structured; nobody samples them; drift is detected by user complaints. Tools include Phoenix, LangSmith, Langfuse, Datadog LLM Observability, and Honeycomb’s LLM tracing,20 each with different trade-offs on framework integration, deployment model, and pricing. EU AI Act Article 72 requires post-market monitoring for high-risk systems; observability infrastructure is the technical layer that satisfies this requirement.28
5. Incident response and rollback. Defined procedure for detecting, triaging, mitigating, and reporting AI incidents; the technical capability to roll back to a previous model version within minutes. Mature: incident response is exercised through tabletop scenarios; rollback is automated and tested; serious incidents are reported to regulators where required (Article 73 of the EU AI Act for high-risk systems; Article 55(c) for GPAI systemic-risk providers; the FDA serious-event reporting framework for medical AI).28 Immature: rollback is a manual procedure that has never been rehearsed; there is no defined incident severity scale.
6. Data and identity controls. Authorization to access training data, retrieved data, tools, and outputs is enforced at the model boundary, integrated with enterprise identity, and tested against documented threat models. Mature: training data has documented provenance and consent basis; retrieval indices are partitioned by tenant or sensitivity; tool calls inherit caller identity; PII detection runs on inputs and outputs; data residency is enforced for regulated jurisdictions. Immature: a service account with broad access is shared across users; PII handling is per-developer judgment; tenants are isolated by prompt rather than by infrastructure. EU AI Act Article 10 requires high-risk systems to use “training, validation and testing data sets [that] shall be subject to data governance and management practices appropriate for the intended purpose of the high-risk AI system.”28 The MITRE ATLAS framework documents the threat surface for adversarial data poisoning, model extraction, and prompt injection across this control plane.27
7. Decommissioning. Defined process for retiring a model from production: notifying dependents, archiving evidence (training data, evaluation results, deployment logs) for the legally required retention period, and confirming that no service still calls the retired endpoint. Mature: decommissioning is a registry-driven workflow with checks; archived evidence is recoverable on demand for audits or litigation; downstream callers fail loudly rather than silently when an endpoint is removed. Immature: old models accumulate in cloud accounts; nobody is sure which are actually in use; legal cannot produce evidence about a system that ran two years ago. EU AI Act Article 20 places corrective-action and information obligations on providers when a high-risk AI system is no longer in conformity; the decommissioning process operationalizes those obligations.28
Figure 4. Governance as delivery: seven components mapped against the AI delivery lifecycle
| Lifecycle stage | Components active |
|---|---|
| Plan / scope | Model registry (entry), evaluation gates (criteria definition), data and identity controls (consent and DPIA) |
| Build | Model registry (versioning), evaluation gates (CI runs), human checkpoints (rubric design) |
| Pre-deploy | Evaluation gates (release runs, adversarial tests), observability (instrumentation verified) |
| Deploy | Evaluation gates (canary), observability (live traces), incident response (on-call active) |
| Operate | Observability and drift, human checkpoints (sampled review), incident response |
| Retire | Decommissioning, model registry (archive) |
Each component is engineering work. The deliverable is code, infrastructure, and integrated process, not a policy memo.
Section 13. Integrating Governance with Delivery
Governance integrated with delivery looks structurally identical to security integrated with delivery, the practice known as DevSecOps over the past decade. The same patterns work; the same anti-patterns fail.
Patterns that work.
Governance as code. Policy is expressed as machine-checkable definitions: evaluation thresholds as YAML, model approval workflows as Argo or GitHub Actions definitions, data classification as metadata attached to tables. Open Policy Agent (OPA), Cedar, and similar policy engines apply at runtime. The principle is that policy should be reviewable, version-controlled, testable, and automatable in the same systems that engineering already uses.
Automated policy enforcement at the boundary. Enforcement runs at API gateways, model-serving layers, and data-access layers, not in policy documents read by humans. NeMo Guardrails, Llama Guard, and equivalent open-source frameworks implement input/output policies as runtime components.19 AWS Bedrock Guardrails, Azure AI Content Safety, and Google Cloud’s Model Armor expose comparable functionality as managed services. Vendor capability claims should be verified against current documentation; the field is moving fast enough that independent verification matters.
Evaluation gates in CI/CD. Pull requests trigger evaluation runs against the CI golden set; promotion requires passing thresholds; production deploys are blocked by registry checks. The Spotify, Uber, and Netflix platform descriptions all integrate evaluation results into deployment metadata, making it impossible to deploy a model that has not passed the defined gates.23
Federated implementation, central oversight. Application teams own their AI systems and run their own evaluation suites; a central platform team owns shared infrastructure (model registry, evaluation tooling, observability) and the policy framework; a small risk function audits adherence and surfaces patterns. This structure is consistent with NIST AI RMF GOVERN 2 (accountability) and ISO/IEC 42001’s leadership and operation clauses.79
Documentation generated from artifacts. Technical documentation under EU AI Act Article 11 and Annex IV, model cards under prevailing transparency conventions, and audit evidence are generated from the registry, evaluation runs, and observability streams, not authored separately. When the system changes, documentation changes automatically.
Patterns that fail.
Late-stage review boards. A committee that meets quarterly to review AI projects after design, build, and most testing is complete is too late to influence design and too disconnected from implementation to verify it. The committee approves what is presented to it, which is the path of least resistance.
Documentation-first compliance. Policies, principles, and attestations authored as standalone documents, with no link to enforcement. The documents accumulate; the enforcement does not.
Tooling without process. Buying an evaluation platform, an observability platform, or a guardrails service without changing the workflow that uses them produces dashboards no one looks at and gates no one enforces.
Process without tooling. Defining policies and procedures that depend on humans to verify compliance manually. Manual verification scales as O(systems × releases × controls) and breaks first.
Single-vendor capture. Outsourcing the entire governance stack to one vendor’s product line creates lock-in and reduces auditability when the vendor’s behavior changes. Mature programs maintain independence between policy definition, enforcement, and evaluation.
The integration target is that, on any given day, a randomly chosen AI system in production should be answerable on the following questions in fewer than ten minutes, without convening a meeting: What model version is deployed? Who owns it? When was it last evaluated, against what dataset, with what results? What policies apply? What incidents have it had? When is it scheduled for re-evaluation or decommissioning? If the answer takes longer, governance has not been delivered; it has been documented.
Section 14. Three Governance Archetypes
The three archetypes below are original to this publication. They describe the dominant patterns observed across enterprise AI programs and predict their failure modes. No archetype is “right”; each succeeds in some contexts and fails in others. Most large organizations exhibit a mix, often inconsistently between business units.
Compliance-Led governance. AI governance is owned by the chief compliance officer, general counsel, or analogous risk function. Activities center on regulatory mapping, attestations, model inventory, and approval gates administered by a central committee. Documentation is the primary deliverable.
Strengths. Strong regulatory legibility, auditors and regulators receive clear artifacts. Clear escalation paths for risk decisions. Effective in heavily regulated sectors (banking, insurance, healthcare) where the failure of compliance is itself the operational risk.
Failure modes. Late-stage review boards (Section 13). Documentation that diverges from operating reality. Engineering circumvention via “shadow AI”, usage of AI tools outside the governed inventory because the governed path is too slow. Inability to keep pace with technical change because technical literacy in the governance function lags by 12-24 months.
Suitable contexts. Regulated industries with mature compliance functions and slow product cycles; AI uses that map cleanly to existing regulatory categories; organizations where regulatory exposure dominates other risk.
Risk-Led governance. AI governance is owned by the CRO or CISO, modeled on existing operational-risk and information-security frameworks. Activities center on threat modeling, control mapping (often to NIST CSF, ISO 27001, or NIST AI RMF), risk registers, and quantitative or semi-quantitative risk scoring.
Strengths. Risk-based prioritization concentrates effort where exposure is highest. Inherits mature security infrastructure (SIEM, SOC, incident response). Integrates well with NIST AI RMF and ISO/IEC 42001 because both frameworks are risk-structured. Effective for adversarial and security-driven failure modes (prompt injection, data exfiltration, model theft).
Failure modes. Risk registers proliferate without operational follow-through. Quantitative scoring overstates precision (probability × impact estimates have wide error bars on novel AI risks). Bias toward security-shaped risks (confidentiality, integrity, availability) and underweighting of fairness, accuracy, and business-outcome risks that are equally important. Reliance on annual or semi-annual assessments rather than continuous monitoring.
Suitable contexts. Organizations with strong security cultures; AI systems that share infrastructure with regulated data; environments where adversarial threats are the dominant risk source.
Engineering-Led governance. AI governance is owned by the engineering function, typically a platform team, ML platform team, or AI platform team, and implemented as code, infrastructure, and CI/CD gates. Activities center on the operational components of Section 12, with policy expressed as machine-checkable definitions, enforcement at the boundary, and documentation generated from artifacts.
Strengths. Governance moves at the speed of delivery. Policies that cannot be enforced in code are exposed as policies that cannot be enforced. Engineering teams own controls rather than circumventing them. Best fit for the operational test in Section 13: any system can be answered for in minutes.
Failure modes. Underweights fairness, ethics, and societal impact unless those are actively scoped in. Risks confusing technical robustness with overall trustworthiness. May be opaque to non-engineering audiences (regulators, customers, board) without dedicated translation. Relies on engineering culture investment that some organizations lack.
Suitable contexts. Software-native organizations; high-velocity AI delivery; organizations whose primary risk is product failure rather than regulatory or adversarial. Frequently the dominant pattern in ML platform teams at large technology companies.
Figure 6. Three governance archetypes, comparison
| Dimension | Compliance-Led | Risk-Led | Engineering-Led |
|---|---|---|---|
| Owner | CCO / GC | CRO / CISO | CTO / VP Eng / Platform |
| Primary deliverable | Documentation, attestations | Risk registers, controls catalog | Code, infrastructure, gates |
| Cadence | Quarterly / annual | Quarterly | Continuous (CI/CD) |
| Primary strength | Regulatory legibility | Risk prioritization | Operational integration |
| Primary failure mode | Documentation drift; shadow AI | Register without follow-through | Underweighting non-technical risk |
| Best fit | Regulated industries | Security-mature orgs | Software-native orgs |
| Worst fit | Fast product cycles | Heavily fairness-driven domains | Heavy disclosure regimes |
Mature organizations rarely use one archetype alone. The pattern that produces the best operational outcomes is Engineering-Led implementation, Risk-Led prioritization, Compliance-Led disclosure, engineering builds the controls, risk decides what to prioritize, compliance handles external interface, with explicit accountability between the three functions. Without explicit handoffs, hybrid approaches degrade into duplicated effort and gaps between layers. Section 16 returns to this point in the combined maturity framework.
Part C, Synthesis
Section 15. How Evaluation and Governance Connect
The two disciplines are usually treated separately. They are not separate. Evaluation gates are governance components. Drift monitoring is governance evidence. Adversarial evaluation is one of the obligations EU AI Act Article 55 places on GPAI systemic-risk providers. The attempt to govern AI without operational evaluation produces documents; the attempt to evaluate AI without governance produces dashboards no one acts on.
The connections are concrete.
Evaluation gates as the implementation of risk management. EU AI Act Article 9 requires a risk management system “established, implemented, documented and maintained” for high-risk AI; Article 15 requires accuracy, robustness, and cybersecurity to be designed in.28 Operationally, these obligations are met by evaluation gates that test for regression, adversarial resistance, and stability, and by the registry that records gate results against versions. NIST AI RMF MEASURE function maps onto evaluation; the GOVERN, MAP, and MANAGE functions map onto the surrounding workflow.7
Adversarial testing as governance and as evaluation. Section 7’s red-team work is simultaneously evaluation (does the system resist attacks?) and governance (does the organization meet its obligation to test, document, and remediate?). MITRE ATLAS provides a shared vocabulary that engineering and audit functions can both reference.27
Observability as the audit trail. The traces, logs, and metrics produced for evaluation and operation are the same artifacts an auditor or regulator wants to see for governance. Building two parallel collection systems wastes effort and produces inconsistencies.
Human-in-the-loop as evaluation and as control. Sampled human review of production outputs serves two purposes: it generates labels for the evaluation flywheel, and it satisfies human-oversight obligations under Article 14.28
Decommissioning as the closing of both loops. When a model is retired, its evaluation history and its governance evidence are archived together. They were never separate.
The interlocking maturity ladder follows from this. An organization cannot have mature evaluation without a model registry; it cannot have a useful model registry without evaluation results to register. It cannot have continuous evaluation without observability; it cannot have observability without governance over the data that flows through it. It cannot have effective adversarial testing without incident response to handle what red-teams find. Each component depends on the others.
The practical consequence is that programs which try to advance one discipline while ignoring the other stall. A team that builds excellent evaluation infrastructure but no model registry cannot answer what is in production. A team that builds excellent registry but no evaluation has a list of unevaluated models. A team that builds both but no observability cannot tell whether the deployed system still matches the evaluated one. The seven operational components of Section 12 are interdependent; advancing them together is the only path that compounds.
This does not mean a program must build everything before delivering value. It does mean that an organization’s overall maturity is bounded by its weakest component. The next section operationalizes that observation as a four-level ladder.
Section 16. A Combined Maturity Framework
The combined maturity framework below is original to this publication. Each level is defined operationally so it can be tested by inspection rather than self-attestation.
Level 1, Ad hoc. AI projects exist; evaluation is whatever the building team chose; governance is reactive (incident → policy memo). Operational test: when asked, the organization cannot list its production AI systems within a working day. Models in production may not have current evaluation results. Incidents may not be detected, or may be detected only by external parties.
Level 2, Defined. A central inventory exists; standard evaluation methods are documented and applied to high-priority systems; governance roles are assigned. Operational test: any production AI system has a named owner, a registry entry, an evaluation result less than six months old, and a documented data classification. Incident response is defined and tested at least annually. Adversarial testing is conducted for customer-facing systems before launch but not continuously. Drift monitoring is partial.
Level 3, Integrated. Evaluation gates run in CI/CD; the model registry is the source of truth referenced in deployment; observability covers all production AI; human-in-the-loop checkpoints have measured override rates; incident response runs on-call. Operational test: a randomly chosen production AI system can be answered on registry, evaluation, ownership, recent incidents, and decommissioning schedule in under ten minutes. Adversarial test sets are versioned and run on every release. Drift triggers automated re-evaluation. Documentation is generated from artifacts.
Level 4, Adaptive. Evaluation, governance, and engineering operate as a single product; failure modes from production feed automatically into golden sets, judge prompts, and adversarial tests; governance changes are deployed through the same CI/CD pipeline as code; the organization participates in industry-wide incident sharing and threat intelligence. Operational test: a regression detected in production leads to a hotfix, a registry update, an evaluation-set extension, and a documentation regeneration within a single working day, without human coordination overhead beyond the normal incident-response process.
Figure 7. Combined maturity ladder
| Level | Evaluation | Governance | Operational test |
|---|---|---|---|
| 1, Ad hoc | Per-team, undocumented | Reactive policy memos | Cannot list production AI in a day |
| 2, Defined | Standard methods on priority systems | Roles assigned; inventory exists | Each system has owner, registry, recent eval |
| 3, Integrated | CI/CD gates; observability; adversarial testing | Registry-driven; documentation generated | Any system answerable in <10 min |
| 4, Adaptive | Continuous flywheel from production | Governance-as-code, deployed with system | Regression → fix → registry → eval → docs in <1 day |
Most enterprises in 2026 are at Level 1 or low Level 2 for AI, even when they are at Level 3 or 4 for traditional software delivery. The capability gap is real and is the operational core of the trust deficit between AI promises and AI delivery.
Section 17. Methodology and Sources
This report was written from a single research perspective, drawing on three categories of evidence with explicit tier discipline.
Tier A, primary sources. Original peer-reviewed papers (RAGAS,1 G-Eval,2 MT-Bench,3 BLEURT,4 BERTScore,5 Greshake et al. on indirect prompt injection,26 Zou et al. on universal adversarial attacks,25 Shankar et al. on validator alignment19); regulatory texts in their published form (Regulation (EU) 2024/1689,28 NIST AI 100-1,7 NIST AI 600-1,8 ISO/IEC 42001:2023,9 EOs 14110, 14148, 1417931); standards-body publications and primary regulatory/judicial records (Moffatt v. Air Canada 2024 BCCRT 149;11 Mata v. Avianca, 678 F. Supp. 3d 443;12 EEOC v. iTutorGroup consent decree13).
Tier B, engineering practice writing. Named-team blog posts and conference materials (Hamel Husain on production evals,1721 Eugene Yan on LLM patterns,18 Uber Engineering on Michelangelo23). Practitioner guidance was used to operationalize abstract principles and was cited where original primary sources would not provide the same operational specificity.
Excluded as load-bearing evidence. Vendor “responsible AI” marketing material; opinion writing without operational substance; secondary-press summaries of incidents where primary records exist; and unverified claims by tools about their own evaluation performance.
Limitations. The legal cases cited are correct as of mid-2026; sanctioning practice is evolving rapidly and case counts (e.g., Charlotin’s database) are moving targets. The EU AI Act’s high-risk implementation guidance under Article 6(5) was due in early 2026 and will likely refine some interpretations cited here. Vendor capability descriptions (LangSmith, Phoenix, MLflow, RAGAS) reflect documentation accessed in late April and early May 2026; readers should verify against current versions before relying on specific feature claims. The three governance archetypes and four-level maturity framework are analytic constructs of this publication; they have not been independently validated and should be treated as a hypothesis to be tested against organizational reality, not as empirical findings. Where primary regulatory interpretation is contested, notably the boundaries of Annex III high-risk classification under the EU AI Act and the definition of “systemic risk” for GPAI models, the report cites multiple authorities rather than asserting a settled reading.
Trust in enterprise AI is not earned by principles or by promises. It is earned by visibly delivered evaluation and visibly delivered governance, every release, on every system, in a way that anyone outside the building team can verify in under ten minutes. The disciplines exist. Most organizations have not yet built them.
- Es, S., James, J., Espinosa-Anke, L., & Schockaert, S. (2024). “Ragas: Automated Evaluation of Retrieval Augmented Generation.” Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics: System Demonstrations, 150-158, St. Julians, Malta. arXiv:2309.15217. https://arxiv.org/abs/2309.15217 ; https://aclanthology.org/2024.eacl-demo.16/ (accessed May 2, 2026). ↩↩↩↩↩
- Liu, Y., Iter, D., Xu, Y., Wang, S., Xu, R., & Zhu, C. (2023). “G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment.” Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2511-2522, Singapore. DOI:10.18653/v1/2023.emnlp-main.153. arXiv:2303.16634. https://aclanthology.org/2023.emnlp-main.153/ (accessed May 2, 2026). ↩↩↩↩
- Zheng, L., Chiang, W.-L., Sheng, Y., Zhuang, S., Wu, Z., Zhuang, Y., Lin, Z., Li, Z., Li, D., Xing, E.P., Zhang, H., Gonzalez, J.E., & Stoica, I. (2023). “Judging LLM-as-a-Judge with MT-Bench and Chatbot Arena.” Advances in Neural Information Processing Systems 36 (NeurIPS 2023). arXiv:2306.05685. https://arxiv.org/abs/2306.05685 (accessed May 2, 2026). ↩↩↩↩↩↩↩↩
- Sellam, T., Das, D., & Parikh, A.P. (2020). “BLEURT: Learning Robust Metrics for Text Generation.” Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 7881-7892. arXiv:2004.04696. https://aclanthology.org/2020.acl-main.704/ (accessed May 2, 2026). ↩↩
- Zhang, T., Kishore, V., Wu, F., Weinberger, K.Q., & Artzi, Y. (2020). “BERTScore: Evaluating Text Generation with BERT.” International Conference on Learning Representations (ICLR 2020). arXiv:1904.09675. https://openreview.net/forum?id=SkeHuCVFDr (accessed May 2, 2026). ↩↩
- Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence (Artificial Intelligence Act). Official Journal of the European Union, L series, 12 July 2024. https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=OJ:L_202401689 (accessed May 2, 2026). ↩
- National Institute of Standards and Technology. (2023). Artificial Intelligence Risk Management Framework (AI RMF 1.0). NIST AI 100-1, January 26, 2023. https://nvlpubs.nist.gov/nistpubs/ai/nist.ai.100-1.pdf ; https://www.nist.gov/itl/ai-risk-management-framework (accessed May 2, 2026). ↩↩↩↩↩↩↩↩↩
- Autio, C., Schwartz, R., Dunietz, J., Jain, S., Stanley, M., Tabassi, E., Hall, P., & Roberts, K. (2024). Artificial Intelligence Risk Management Framework: Generative Artificial Intelligence Profile. NIST AI 600-1, July 26, 2024. DOI:10.6028/NIST.AI.600-1. https://nvlpubs.nist.gov/nistpubs/ai/NIST.AI.600-1.pdf (accessed May 2, 2026). ↩↩↩↩
- ISO/IEC 42001:2023. Information technology, Artificial intelligence, Management system. International Organization for Standardization, December 2023. https://www.iso.org/standard/42001 (accessed May 2, 2026). ↩↩↩↩↩↩
- Executive Order 14179 of January 23, 2025, “Removing Barriers to American Leadership in Artificial Intelligence.” 90 Fed. Reg. 8741. https://www.federalregister.gov/documents/2025/01/31/2025-02172 (accessed May 2, 2026). ↩
- Moffatt v. Air Canada, 2024 BCCRT 149 (B.C. Civil Resolution Tribunal, February 14, 2024). Tribunal Member Christopher C. Rivers. https://www.canlii.org/en/bc/bccrt/doc/2024/2024bccrt149/2024bccrt149.html (accessed May 2, 2026). ↩↩↩↩
- Mata v. Avianca, Inc., 678 F. Supp. 3d 443 (S.D.N.Y. 2023). Opinion and Order on Sanctions, Judge P. Kevin Castel, June 22, 2023. ECF No. 54. https://law.justia.com/cases/federal/district-courts/new-york/nysdce/1:2022cv01461/575368/54/ (accessed May 2, 2026). ↩↩↩↩↩↩↩
- EEOC v. iTutorGroup, Inc., et al., No. 1:22-cv-02565-PKC-PK (E.D.N.Y., Consent Decree filed Aug. 9, 2023; approved September 8, 2023). U.S. Equal Employment Opportunity Commission press release, August 9, 2023. https://www.eeoc.gov/newsroom/itutorgroup-pay-365000-settle-eeoc-discriminatory-hiring-suit (accessed May 2, 2026). ↩↩↩↩
- Lacey v. State Farm General Insurance Co. (C.D. Cal. 2025), reported sanctions order. Coverage and the Damien Charlotin database of AI-related judicial reprimands referenced via Esquire Deposition Solutions, “Federal Court Turns Up the Heat on Attorneys Using ChatGPT for Research,” 2025. https://www.esquiresolutions.com/federal-court-turns-up-the-heat-on-attorneys-using-chatgpt-for-research/ (accessed May 2, 2026). ↩↩↩↩
- DPD chatbot incident, January 18, 2024. DPD statement to media including BBC and TIME. TIME, “AI Chatbot Curses at Customer and Criticizes Work Company,” January 20, 2024. https://time.com/6564726/ai-chatbot-dpd-curses-criticizes-company/ ; OECD AI Incidents Monitor (Incident 631), https://incidentdatabase.ai/cite/631/ (accessed May 2, 2026). ↩↩↩
- Chevrolet of Watsonville chatbot incident, December 17-18, 2023. Documented via Chris Bakke’s X post (now archived) and contemporaneous coverage including GM Authority, “GM Dealer Chat Bot Agrees To Sell 2024 Chevy Tahoe For $1.” https://gmauthority.com/blog/2023/12/gm-dealer-chat-bot-agrees-to-sell-2024-chevy-tahoe-for-1/ (accessed May 2, 2026). ↩↩↩
- Husain, H. (2024). “Your AI Product Needs Evals.” Hamel’s Blog. https://hamel.dev/blog/posts/evals/ (accessed May 2, 2026). ↩↩↩↩↩
- Yan, E. (2023). “Patterns for Building LLM-based Systems & Products.” eugeneyan.com, July 2023. https://eugeneyan.com/writing/llm-patterns/ (accessed May 2, 2026). See also Yan, E., Bischof, B., Frye, C., Husain, H., Liu, J., & Shankar, S. (2024). “What We Learned from a Year of Building with LLMs (Part II).” O’Reilly Radar, May 31, 2024. https://www.oreilly.com/radar/what-we-learned-from-a-year-of-building-with-llms-part-ii/ (accessed May 2, 2026). ↩↩↩
- Shankar, S., Zamfirescu-Pereira, J.D., Hartmann, B., Parameswaran, A.G., & Arawjo, I. (2024). “Who Validates the Validators? Aligning LLM-Assisted Evaluation of LLM Outputs with Human Preferences.” Proceedings of the 37th Annual ACM Symposium on User Interface Software and Technology (UIST ‘24), October 13-16, 2024, Pittsburgh, PA. arXiv:2404.12272. DOI:10.1145/3654777.3676450. https://arxiv.org/abs/2404.12272 (accessed May 2, 2026). ↩↩↩↩↩↩
- LangChain. LangSmith Documentation, accessed May 2, 2026. https://docs.smith.langchain.com . Arize AI. Phoenix Documentation, accessed May 2, 2026. https://arize.com/docs/phoenix/ . Langfuse documentation, https://langfuse.com . Comparative analyses: Arize, “Comparing LLM Evaluation Platforms” (https://arize.com/llm-evaluation-platforms-top-frameworks/) and Digital Applied, “Agent Observability: LangSmith, Langfuse, Arize 2026” (https://www.digitalapplied.com/blog/agent-observability-platforms-langsmith-langfuse-arize-2026). Vendor capability claims neutrally summarized; readers should verify against current documentation. ↩↩↩↩↩↩
- Husain, H. (2025). “LLM Evals: Everything You Need to Know.” Hamel’s Blog. https://hamel.dev/blog/posts/evals-faq/ (accessed May 2, 2026). ↩↩↩↩↩
- Mikulski, B. “Shadow deployment vs. canary release of machine learning models.” https://mikulskibartosz.name/shadow-deployment-vs-canary-release ; reproduced at JFrog ML https://www.qwak.com/post/shadow-deployment-vs-canary-release-of-machine-learning-models (accessed May 2, 2026). ↩
- Hermann, J., & Del Balso, M. (2017). “Meet Michelangelo: Uber’s Machine Learning Platform.” Uber Engineering Blog, September 5, 2017. https://www.uber.com/blog/michelangelo-machine-learning-platform/ . Wang, K., Cai, M., Wang, J., & Chen, E. (2024). “From Predictive to Generative, How Michelangelo Accelerates Uber’s AI Journey.” Uber Engineering Blog. https://www.uber.com/blog/from-predictive-to-generative-ai/ (accessed May 2, 2026). ↩↩↩↩↩↩↩
- Pan, T. (2026). “Releasing AI Features Without Breaking Production: Shadow Mode, Canary Deployments, and A/B Testing for LLMs.” TianPan.co, April 9, 2026. https://tianpan.co/blog/2026-04-09-llm-gradual-rollout-shadow-canary-ab-testing (accessed May 2, 2026). ↩
- Zou, A., Wang, Z., Carlini, N., Nasr, M., Kolter, J.Z., & Fredrikson, M. (2023). “Universal and Transferable Adversarial Attacks on Aligned Language Models.” arXiv:2307.15043. https://arxiv.org/abs/2307.15043 (accessed May 2, 2026). See also Weng, L. (2023). “Adversarial Attacks on LLMs,” https://lilianweng.github.io/posts/2023-10-25-adv-attack-llm/ . ↩↩↩
- Greshake, K., Abdelnabi, S., Mishra, S., Endres, C., Holz, T., & Fritz, M. (2023). “Not what you’ve signed up for: Compromising Real-World LLM-Integrated Applications with Indirect Prompt Injection.” Proceedings of the 16th ACM Workshop on Artificial Intelligence and Security (AISec ‘23), Copenhagen, November 30, 2023, pp. 79-90. DOI:10.1145/3605764.3623985. arXiv:2302.12173. https://arxiv.org/abs/2302.12173 (accessed May 2, 2026). ↩↩
- MITRE Corporation. Adversarial Threat Landscape for AI Systems (ATLAS). v5.1.0, November 2025. https://atlas.mitre.org/ ; ATLAS fact sheet at https://atlas.mitre.org/pdf-files/MITRE_ATLAS_Fact_Sheet.pdf (accessed May 2, 2026). ↩↩↩
- Regulation (EU) 2024/1689 (full text, including Articles 5, 6, 9-15, 16-22, 26, 49, 50, 51-55, 72, 73, 99, 113; Annexes I, III, IV, XI-XIII). Official Journal of the European Union, L series, 12 July 2024. https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=OJ:L_202401689 . Article-level summaries used for cross-reference: https://artificialintelligenceact.eu (accessed May 2, 2026). European Commission. (2025). Guidelines on the scope of obligations for providers of general-purpose AI models. July 18, 2025. https://digital-strategy.ec.europa.eu/en/policies/contents-code-gpai (accessed May 2, 2026). ↩↩↩↩↩↩↩↩↩↩↩↩↩↩↩↩↩
- Regulation (EU) 2024/1689, Article 99 (Penalties); Article 100 (Administrative fines on Union institutions); Article 101 (Fines for providers of general-purpose AI models). https://artificialintelligenceact.eu/article/99/ (accessed May 2, 2026). ↩
- Soler Garrido, J., et al. (2024), “Interplay of ISMS and AIMS in context of the EU AI Act.” arXiv:2412.18670. https://arxiv.org/abs/2412.18670 (accessed May 2, 2026). ↩
- Executive Order 14110 of October 30, 2023, “Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence.” 88 Fed. Reg. 75191. Executive Order 14148 of January 20, 2025, “Initial Rescissions of Harmful Executive Orders and Actions.” Executive Order 14179 of January 23, 2025, “Removing Barriers to American Leadership in Artificial Intelligence.” 90 Fed. Reg. 8741. The White House. https://www.whitehouse.gov/presidential-actions/2025/01/removing-barriers-to-american-leadership-in-artificial-intelligence/ ; Federal Register, https://www.federalregister.gov/documents/2025/01/31/2025-02172 (accessed May 2, 2026). ↩↩↩
- New York City Local Law 144 of 2021 (codified as N.Y.C. Admin. Code §§ 20-870-20-874). Final Rule, 6 RCNY §§ 5-300-5-304, NYC Department of Consumer and Worker Protection, April 6, 2023; enforcement effective July 5, 2023. https://www.nyc.gov/site/dca/about/automated-employment-decision-tools.page (accessed May 2, 2026). NY State Comptroller audit, “Enforcement of Local Law 144, Automated Employment Decision Tools,” December 2, 2025, https://www.osc.ny.gov/state-agencies/audits/2025/12/02/enforcement-local-law-144-automated-employment-decision-tools (accessed May 2, 2026). ↩
- Linux Foundation / Databricks. MLflow Documentation, version 2.x, accessed May 2, 2026. https://mlflow.org/docs/latest/ml/model-registry . Databricks Unity Catalog Model Registry documentation, https://docs.databricks.com/aws/en/mlflow/ (accessed May 2, 2026). ↩
- Wataoka, K., et al. (2024). “Self-Preference Bias in LLM-as-a-Judge.” arXiv:2410.21819. https://arxiv.org/abs/2410.21819 (accessed May 2, 2026). ↩
Was this useful?
Related
Membership
Become a Member to receive new research as they are published.
