Most AI decision engine deployments that fail do not fail because of the model. They fail because one of six conditions was not in place before build began — and no one checked. This framework is a structured assessment of those six conditions. Work through each gate before committing to a build timeline. Any gate that scores zero should stop the project until it is resolved.
What this framework assesses
The AI Decision Engine Readiness Framework evaluates the six conditions that consistently determine whether a production AI decision engine delivers durable business value. These conditions are derived from the failure pattern documented in Why Most AI Projects Fail — each gate corresponds directly to a root cause of production failure.
The framework is designed to be run before build begins, not during build and certainly not after launch. It takes approximately 60–90 minutes to assess honestly with the relevant stakeholders in the room. The output is a readiness score and a gap list that defines what must be resolved before a realistic production timeline can be set.
How to score each gate
Score each gate on a 0–2 scale:
- 0 — Not ready. The condition is not met and no clear path exists to meeting it. Build should not proceed until this gate is resolved.
- 1 — Partial. The condition is partially met. There are specific gaps that are addressable within the project timeline but must be tracked explicitly as dependencies.
- 2 — Ready. The condition is fully met. The project can proceed past this gate without remediation work.
Gate 1: Decision Definition
Is the decision type fully defined?
The decision being automated must be clearly described before any build work begins. Vague decisions — "help sales prioritise better" or "automate compliance review" — are not decisions. They are problem statements that still need to be translated into a specific, recurring judgment with defined inputs, explicit criteria, and a measurable outcome.
Score 2 if all four conditions are true:
- The decision type is recurring — the same judgment is made repeatedly across a defined workflow.
- The criteria are explicit — you can describe the logic as "if X and Y then Z" without relying on unstated institutional knowledge.
- A named KPI owner exists — one person is accountable for the business metric this decision affects post-launch.
- Success is measurable — there is a specific operational metric (conversion rate, resolution time, escalation rate, error rate) that will move if the decision engine performs correctly.
Score 1 if:
- The decision is described but the criteria have not been made fully explicit, or the KPI exists but has no named owner.
Score 0 if:
- The decision type is still described as a problem area rather than a specific recurring judgment, or no one can articulate what operational metric will improve.
Gate 2: Data Readiness
Is the operational data complete and consistent at inference time?
The decision engine needs a defined set of data fields to reason about each decision. Those fields must be present, consistent, and current when the decision run is triggered — not available in theory from a data warehouse query, but retrievable in practice from the operational system at the moment the trigger fires.
Score 2 if all four conditions are true:
- The required data fields are defined — you have a canonical entity schema listing every field the engine needs.
- Population rate is sufficient — for each required field, the actual population rate in the operational system meets a defined minimum threshold (typically 85%+ for core fields).
- Consistency is acceptable — the same field means the same thing across different teams, regions, and entry points that feed the system.
- A fallback is defined — there is a specified behaviour for every field that may be missing at inference time (use default, skip, escalate, or hold).
Score 1 if:
- Core fields are available but some secondary fields have inconsistent population, or the fallback for missing fields has not been defined.
Score 0 if:
- The data required for the decision has not been audited, or known quality issues exist in core fields with no remediation plan.
Gate 3: Knowledge Readiness
Is company knowledge structured, current, and retrievable?
An AI decision engine reasons against company-specific knowledge — policy documents, eligibility rules, product guidelines, regulatory constraints, approval authority matrices. That knowledge must be available in a form the engine can retrieve at inference time. Documents that exist on a SharePoint drive but have not been indexed, validated, and structured for retrieval are not usable knowledge.
Score 2 if all four conditions are true:
- The knowledge sources required for the decision are identified — you have a list of every policy document, rule set, or guideline the engine needs to reason correctly.
- Sources are current — each source has a confirmed last-updated date and an owner responsible for keeping it accurate.
- Sources are structured for retrieval — documents have been cleaned, chunked, and are ready for indexing into a vector store, or this work is scoped with a defined completion date.
- An update process is defined — there is a mechanism for propagating policy changes into the knowledge base when they occur, not on an ad-hoc basis.
Score 1 if:
- Sources are identified and reasonably current but have not been structured for retrieval, or no update process exists yet.
Score 0 if:
- The relevant knowledge sources have not been identified, or known sources are significantly out of date with no owner.
Gate 4: Architecture Readiness
Does the planned system include production controls?
A production AI decision engine is not a model plus an API. It is a system with controls: confidence gating, external policy checks, structured output, escalation routing, and audit logging. These cannot be added as an afterthought once the model layer is built. They must be part of the architecture specification from the start, because the model layer is designed around them.
Score 2 if all five conditions are true:
- A confidence threshold mechanism is defined — there is a specified threshold per decision type that gates automated action versus human escalation.
- Policy checks are externalised — the compliance rules that must be enforced regardless of model output are identified and will be implemented as hardcoded checks outside the model.
- The output schema is defined — the structure of every decision run output (decision type, confidence score, rationale, policy checks, action triggered) is specified before build begins.
- An escalation path is designed — low-confidence and policy-blocked decisions have a defined route to a human reviewer, with specified context delivered to the reviewer.
- An audit log schema is defined — the fields that every decision run record must contain are specified and will be implemented from day one.
Score 1 if:
- Some control layers are defined but others (typically audit log or escalation path) are deferred to a later phase.
Score 0 if:
- The architecture plan does not yet include a confidence evaluation layer, policy checks, or audit logging — the current plan is model input/output without production controls.
Gate 5: Workflow Integration
Is there a defined path from decision output to workflow action?
A decision engine that delivers output to a human who then manually acts on it is not a decision engine — it is a recommendation tool. For the system to deliver operational leverage, the decision output must trigger a workflow action in an operational system automatically. This requires API access, schema mapping, write-back logic, and error handling. None of these are trivial, and all of them must be scoped before build begins.
Score 2 if all four conditions are true:
- The receiving system is identified — the specific operational system (CRM, case management tool, ERP, task queue) that will receive the decision output is named.
- API or integration access is confirmed — the technical mechanism for the write-back is confirmed with IT or the system owner, not assumed.
- The triggered action is defined — the specific workflow state change or task that the decision output creates is described (e.g., "CRM record status updated to Qualified, task assigned to rep X, priority set to High").
- Error handling is scoped — the behaviour when the write-back fails is defined (retry logic, fallback notification, decision record preserved).
Score 1 if:
- The receiving system is identified and access is likely, but integration has not been confirmed with the system owner, or error handling has not been scoped.
Score 0 if:
- The integration path has not been defined — the plan is to deliver output to a dashboard or email and have humans act on it, with no live workflow write-back.
Gate 6: Governance Readiness
Is post-launch monitoring, escalation review, and ownership defined?
Most AI projects are scoped to launch. Governance — the ongoing monitoring, escalation review, knowledge maintenance, and performance accountability that keeps the system performing after launch — is not scoped or funded. This is the single most common reason a system that works at launch degrades within six months. Governance must be defined before go-live, not after the first quality problem is discovered.
Score 2 if all four conditions are true:
- Business outcome monitoring is defined — the KPI metric from Gate 1 has a review cadence and a named reviewer who will assess whether it is moving in the right direction.
- Output quality monitoring is defined — decision type distribution and escalation rate are tracked and have alert thresholds that trigger investigation.
- Knowledge update ownership is assigned — every knowledge source in Gate 3 has a named owner with a defined update schedule and a process for propagating changes into the system.
- Escalation review is operational — the process for a human reviewer to receive an escalated decision, assess it, take action, and provide structured feedback to the system is defined and assigned.
Score 1 if:
- Monitoring intent exists but specific metrics, cadences, or owners are not yet assigned — it is on the plan but not confirmed.
Score 0 if:
- Post-launch governance has not been scoped. The project plan ends at go-live with no monitoring, escalation review process, or ownership defined for ongoing performance.
Scoring your readiness
Add your scores across all six gates. Maximum score is 12.
All six conditions are substantially in place. A realistic production timeline can be set and build can begin. Use the AI Opportunity Diagnostic to validate architecture decisions and integration path before committing to a vendor or build approach.
One or two gates are scoring 0 or 1. Build can begin on the gates that are ready, but the gaps in the lower-scoring gates must be resolved in parallel and treated as hard dependencies before production deployment. Each gate scoring below 2 requires a specific remediation plan with a deadline and owner. The AI Opportunity Diagnostic is designed to map this remediation path.
Multiple foundational conditions are not in place. Starting a build in this state has a high probability of producing a pilot that cannot be taken to production. The right move is to spend 4–8 weeks resolving the Gate 1 and Gate 2 conditions first — decision definition, data readiness, knowledge inventory — before any model or architecture work begins. The AI Opportunity Diagnostic identifies exactly which conditions to prioritise and in what order.
What to do with a gap
Every gate that scores below 2 generates a specific remediation task. The task is not "do more research" — it is a concrete action with an owner and a deadline:
- Gate 1 gap: Schedule a decision definition workshop with the operational owner. Produce a one-page decision specification: trigger, inputs, criteria, output format, KPI, and KPI owner. This document becomes the acceptance criterion for the project.
- Gate 2 gap: Run a data quality audit on each required field in the operational system. Produce a field-level assessment with population rates and consistency scores. Define fallback behaviour for every field below threshold before build begins.
- Gate 3 gap: Audit every knowledge source required for the decision. Assign a named owner to each source. Define the update schedule and the process for propagating changes into the retrieval system. Complete the structuring and indexing work as a project phase before the reasoning layer is built.
- Gate 4 gap: Produce a production architecture specification that includes confidence threshold definition, externalised policy check list, output schema, escalation path design, and audit log schema. This document should be reviewed and approved before build begins, not produced during build.
- Gate 5 gap: Map the integration requirements. Confirm API or webhook access with the receiving system's owner. Define the exact workflow state change the decision output triggers. Scope the error handling. Include integration timeline in the project plan as a dependency, not an assumption.
- Gate 6 gap: Produce a governance specification: monitoring metrics and cadence, alert thresholds, escalation review process, knowledge update ownership, and a named reviewer for post-launch KPI accountability. Make this a deliverable of the project, due before go-live.
Why this framework exists
This framework was developed from direct observation of where AI decision engine projects break down. The six gates map exactly to the five failure modes documented in Why Most AI Projects Fail — Gate 4 splits architecture into its own gate because it is the most commonly underspecified condition and the most expensive to fix retroactively.
The framework is not a checklist for its own sake. It is a forcing function for having the right conversations before build begins — conversations that most AI projects defer until they become blockers. Every condition in the framework is knowable before a line of code is written. Projects that assess these conditions honestly before committing to a build timeline fail at a fraction of the rate of projects that do not.
If your assessment reveals gaps across multiple gates and you want a structured path through the remediation, that is the purpose of the AI Opportunity Diagnostic. One session. Specific outputs. No commitment to a build approach required.
Start Your AI Opportunity Diagnostic
Read about the AI Decision Engine Architect engagement model →
Related Framework Pages
- AI Implementation Roadmap
- AI Governance Framework
- AI Data Strategy Framework
- AI Deployment Framework