Separating Logic from Models: Why Your AI System Needs a Decision Plane

AI systems that embed decision logic inside prompts are fast to build and impossible to govern. The decision plane concept offers a better architectural model — learn what it is, why it matters, and how to implement it.

Separating Logic from Models: Why Your AI System Needs a Decision Plane

When teams build AI systems, they typically put decision logic inside prompts or model calls. The instruction says "if the customer's account is past due, do not offer a discount." The logic is there — technically. It works in demos. It fails in production.

It fails because a prompt is a suggestion, not a constraint. A model is a probabilistic function, not a policy enforcement point. When consequential logic — the kind that affects customers, triggers financial transactions, or determines access to services — lives inside a model call, it is invisible, untestable, and impossible to roll back independently of the application. The decision plane concept exists to fix this architectural error.

This article explains what a decision plane is, where the concept comes from, what it looks like in practice, and why every AI system that takes consequential actions needs one.

The Analogy: Network Planes and What They Teach AI Architects

The decision plane concept borrows from a well-established pattern in network architecture: the separation of planes. In traditional networking, a router or switch operates across three distinct planes:

  • Data plane: moves packets from one interface to another; it executes
  • Control plane: computes routing tables and determines where packets should go; it decides
  • Management plane: handles configuration, monitoring, and policy; it governs

These planes are separated deliberately. The data plane is optimized for speed — it needs to forward millions of packets per second without thinking about routing policy. The control plane is optimized for correctness — it builds the routing table using algorithms that can be updated, tested, and replaced without touching data plane hardware. Mixing them produces systems that are fast but ungovernable, or governable but slow.

The same principle applies to AI systems. The model is your data plane: fast, powerful, stateless, and best understood as an execution engine that processes inputs and generates outputs. The decision plane is your control plane: the layer where logic about what the system should do in a given situation is encoded, tested, versioned, and enforced.

The arXiv survey on agentic AI architectures makes this separation explicit: the most robust agent architectures separate the reasoning layer (what the model does) from the execution layer (what the system does) and from the governance layer (what the system is permitted to do). The decision plane is the governance layer made concrete.

Three Architectural Anti-Patterns

Before describing what the decision plane is, it is useful to describe what it is not — by cataloging the three anti-patterns most commonly found in AI systems that lack one.

Anti-Pattern 1: Logic in Prompts

This is the most common pattern and the most fragile. The system prompt includes instructions like "never offer refunds to users who have been active for fewer than 30 days" or "only escalate to a human if confidence is below 0.7." These instructions are processed by the model on every inference call.

The failure modes are numerous. Models do not reliably follow complex conditional instructions, especially when those instructions conflict with the goal expressed elsewhere in the prompt. The logic cannot be tested independently of the model — you must run the full model call to verify that a rule is being applied. It cannot be versioned separately from the application. And it provides no audit trail: there is no record of which rule was evaluated and what its outcome was at the time of a specific decision.

Anthropic's research on building effective agents explicitly cautions against using the model as the sole decision engine for consequential logic. The model is a generator; the decision about whether to act on a generated output belongs in a separate layer.

Anti-Pattern 2: Logic in Orchestration Code

This pattern is more sophisticated but has its own failure modes. The orchestration code — the Python or TypeScript layer that coordinates model calls, tool invocations, and state management — contains if/else logic that governs what the agent can do. The logic is more reliable than prompt logic, but it inherits all the problems of application code: it requires a deploy to change, it is not readable by non-engineers, and it cannot be tested independently of the application.

More subtly, orchestration logic tends to accumulate. What starts as three clean conditional checks becomes, over eighteen months, a sprawling set of nested conditions that no single person fully understands. The business intent behind each condition is lost. The system behaves correctly — until it doesn't — and the post-mortem cannot determine which condition failed or why.

Anti-Pattern 3: Logic in Embeddings

The most subtle anti-pattern embeds decision logic in the retrieval layer. The system retrieves context that contains business rules — "customers on the Starter plan do not have access to the analytics export" — and relies on the model to apply those rules correctly from retrieved context. This conflates facts (what is stored in the knowledge base) with rules (what the system should enforce) and produces a system where decision logic can be silently overridden by retrieval results that contain conflicting information.

Sapiens' analysis of decisioning layers distinguishes cleanly between "data that informs a decision" and "logic that governs a decision." Embedding rules in a knowledge base treats them as data — subject to retrieval variance, chunking artifacts, and model interpretation. A decision plane treats them as logic — deterministic, versioned, and enforced independently of what the model retrieves.

The Decision Plane Defined: Properties of a Proper Implementation

A decision plane is an architectural layer where all logic that could affect a consequential outcome lives in an explicit, testable, and versionable form, separate from the model. It has four defining properties:

1. Externality

Decision logic lives outside the model and outside the orchestration code. It is stored in a dedicated system — a policy engine, a rules repository, or a purpose-built decision platform — that can be read, tested, and modified independently of the application.

2. Explicitness

Every rule is named and its conditions are fully specified. There is no implicit logic, no "the model knows to do X in situation Y." If the rule exists, it is written down. If it is written down, it can be read by any stakeholder with access to the decision plane.

3. Versionability

Every rule has a version history. Changes are tracked. At any point in time, it is possible to retrieve the specific version of a rule that was active during a past decision. This property is what makes a decision plane defensible in regulatory and audit contexts.

4. Testability

Rules can be tested in isolation, without running the full application stack. A product manager can write a test case — "given a Starter-tier account requesting an analytics export, the rule should deny access" — and verify it without involving an engineer or running a model call.

TEKsystems' three-layer safety architecture for enterprise AI formalizes this property: each layer of the architecture must be independently testable, and the decision layer specifically must be verifiable against defined test cases before any rule change reaches production.

Why This Matters Operationally

The decision plane is not an abstract architectural ideal. It has concrete operational consequences across four dimensions.

Testability

With a decision plane, you can write a test suite for your business rules the same way you write a test suite for your application code. Before a rule change goes to production, you run the test suite and verify that the change behaves as intended. This is not possible when your rules live in prompts or orchestration code.

Rollback

When a rule change produces unexpected behavior in production, you can roll back the rule independently of the application. Reverting from rule version 3 to version 2 takes seconds and does not require a deploy. This matters because rule changes often affect customer-facing behavior in ways that need to be reversed quickly — and quickly means without waiting for an engineering cycle.

Auditability

Every decision made by the system can be explained by pointing to the specific rule version that was active at the time, the facts that were evaluated, and the outcome that was produced. This is the property that makes AI systems defensible in regulatory contexts. Without a decision plane, the audit answer to "why did your system do that?" is "the model decided." With a decision plane, the answer is "rule feature-export-tier-gate-v3 evaluated the account's plan tier as Starter and returned deny."

Ownership

When logic lives in the decision plane, it can have a human owner who is not an engineer. Revenue Operations can own the trial conversion rules. Legal can own the compliance rules. Finance can own the refund rules. Each owner can read, test, and propose changes to the rules they own without touching application code. This is the organizational benefit that the technical architecture enables.

A Minimal Implementation Sketch

A minimal decision plane does not require a sophisticated platform. At its simplest, it consists of three components:

  1. A rule repository: a structured store of named, versioned rules. Each rule specifies its conditions (as typed facts) and its outcome (allow, deny, allow-with-constraints). A YAML file in a version-controlled repository is a valid starting point. A purpose-built policy engine like AWS Cedar or Open Policy Agent is the production-grade version.
  2. An evaluation function: a function that takes a set of input facts and a rule name, evaluates the rule against the facts, and returns an outcome. This function must be callable synchronously from your orchestration layer, before any consequential action executes.
  3. An enforcement point: the place in your orchestration code where the evaluation function is called, and where the action is either permitted or blocked based on the outcome. Every consequential action in your system must pass through an enforcement point. Actions that bypass enforcement are outside the decision plane and therefore ungoverned.

This is the minimal structure. From here, you add: a decision log that captures every evaluation event, a test harness for running rule test cases before promotion, and a rollout mechanism that allows rules to be promoted through draft, shadow, and canary stages before going fully active.

For a deeper look at the infrastructure that supports this architecture, see our article on infrastructure for deterministic AI decisions.

The Common Objection: "But Our LLM Is Smart Enough"

The most frequent pushback against decision plane architecture is a form of model confidence: "our model is good enough to follow the instructions we give it; we don't need a separate layer." This objection misunderstands what the decision plane is for.

The decision plane is not a workaround for model limitations. It is not there because you don't trust the model. It is there because consequential decisions need properties — versionability, testability, auditability, independent rollback — that no model can provide, regardless of how capable it is.

A very capable model that correctly follows instructions embedded in a prompt is still producing unversioned, untestable, unauditable decisions. When a regulator asks "show me the rule that governed this decision and prove it was active at the time," a capable model is not helpful. A versioned rule in a decision plane is.

The question is not whether your model is smart enough. The question is whether your system is governable. Those are different questions with different answers.

Building Toward a Decision Plane

If your AI system currently puts decision logic in prompts, orchestration code, or embeddings, the path toward a decision plane does not require rewriting the application. It requires identifying the three to five most consequential decision points in your current system — the actions where a wrong outcome would be most damaging — and building an enforcement point and rule definition for each of them.

Start there. Write the rules down, explicitly, in a format that can be version-controlled. Build the evaluation function. Add the enforcement point. Log every evaluation event. You now have the beginning of a decision plane — and you have made those three to five most consequential decisions visible, testable, and auditable for the first time.

From that foundation, expand. Add rules for the next tier of consequential actions. Build the test harness. Implement staged rollout. Assign rule owners. The decision plane grows with the system, and the governance properties it provides grow with it.

The architecture is not complex. What makes it hard is the discipline required to enforce the principle consistently: every consequential action goes through the plane. No exceptions, no bypasses. That discipline is the difference between a system that is governable and one that merely looks like it might be.

Explore Memrail's Context Engineering Solution

References & Citations

  1. Building Effective Agents (Anthropic)

    Anthropic research on agent architecture patterns, including the risks of using the model as the sole decision engine.

  2. Architectures for Building Agentic AI Systems (arXiv 2512.09458) (arXiv)

    Comprehensive survey of architectural patterns for agentic AI, including the separation of reasoning from execution layers.

  3. Safe AI Implementation: Three-Layer Architecture (TEKsystems)

    Practitioner guidance on implementing a three-layer safety architecture for enterprise AI systems in production.

  4. Show Me Your Decisioning Layer (Sapiens)

    Industry analysis treating the decisioning layer as a managed architectural artifact, separate from model inference.