[THOUGHT LEADERSHIP]

[

12/29/25

]

The Probabilistic Machine and the Deterministic System

[Author]:

Amjad Fatmi

There is a tension at the center of every AI agent deployment that most teams have not named yet. Once you see it, you cannot unsee it.

The machine making decisions is probabilistic. The systems it operates on are deterministic. And nothing in between is designed to bridge that gap.

Two things that do not belong together

A language model does not compute answers. It samples them. Given a prompt, it generates a probability distribution over possible next tokens and draws from it. This is true for every output it produces, including tool calls. When an agent decides to send an email, delete a record, or call an API, that decision is a sample from a distribution. Run the same context twice and you may get different tool calls. The model is not certain. It is likely.

The systems the agent operates on have no concept of likelihood. A database executes the query or it does not. A payment processor charges the card or it does not. An email server sends the message or it does not. These systems are designed around exactness. The same input always produces the same output. There is no "probably deleted." There is deleted.

This is the mismatch. One side operates in probability space. The other operates in binary space. And when the probabilistic machine issues an instruction, the deterministic system executes it with complete certainty, regardless of how uncertain the model was when it generated it.





The model's uncertainty does not transfer downstream. The consequence is exact even when the reasoning was not.

When this did not matter

For most of the history of machine learning, this mismatch was a non-issue. Models produced outputs for humans to evaluate. A recommendation engine suggested products. A classifier flagged emails. A language model drafted text. In every case, a human stood between the probabilistic output and any action taken in the world.

That human was performing a function nobody described explicitly: converting probabilistic outputs into deliberate, reviewed decisions before they reached deterministic systems. The model was uncertain. The human absorbed that uncertainty, applied judgment, and only then issued an instruction.

This is why AI-assisted work felt safe before the question of safety was even on anyone's mind. The architecture was safe by default because a reasoning entity with accountability was always at the boundary.





Agents remove the human from that position. The model's output connects directly to the system's input. The conversion layer is gone. And nothing has been put in its place.


What actually goes wrong

The failure modes that emerge from this architecture are not exotic. They follow directly from the mismatch.

Confident wrongness. A model can produce a tool call with complete syntactic fluency while being wrong about what it should be. The confidence visible in the output format, a well-formed JSON object with correct field names is not evidence of semantic correctness. The model generated what it thought was likely given the context. The deterministic system has no way to distinguish "confidently correct" from "confidently wrong." It executes both.

Accumulated drift. In long agentic tasks, the model's internal picture of the world diverges from reality. The model builds a representation of what has happened based on what it has seen in context. Actual system state changes with each action and does not update the model's representation except through explicit observation. By step twelve of a fifteen-step task, the model may be operating on a picture of the world that has not been true since step four. Every action it takes from that point is a deterministic consequence of a probabilistic misunderstanding.

Amplified edge cases. Models perform well on the center of the distribution and unpredictably on the edges. In production, edge cases are guaranteed to appear. An unusual order, an unexpected parameter combination, a request phrased in an uncommon way. On these inputs, the model's output is less reliable exactly when the stakes of a wrong action may be highest. The deterministic system does not know it is in an edge case. It executes.

The asymmetry nobody talks about

There is an asymmetry in the cost of errors that changes everything about how this system should be designed.

When a model is wrong and a human catches it before any action is taken, the cost is a few seconds of review. When a model is wrong and an agent acts on it before anyone notices, the cost is a real consequence in a real system. A refund that should not have been issued. An email sent to the wrong recipient. An infrastructure change that took down a service. A record that cannot be undeleted.

The cost of catching the error is always small. The cost of not catching it varies from minor to catastrophic. This asymmetry is the entire argument for a mandatory evaluation layer between model output and system execution.

Most agent deployments today are designed as if the costs are symmetric. As if letting the model output reach the system directly is approximately as safe as reviewing it first. It is not. The costs are not symmetric. The consequences of being wrong are often irreversible. The cost of requiring explicit authorization before execution is always recoverable.





Why probabilistic governance does not fix this

The instinct when people recognize this problem is to add more machine reasoning to the solution. Train a classifier to detect bad actions. Use another model to evaluate the first model's outputs. Build a guardrail that scores outputs before they execute.

These approaches are probabilistic. They add a probabilistic layer to evaluate a probabilistic output before it reaches a deterministic system. The fundamental problem is not solved. It is compounded.

A probabilistic evaluator has its own confidence distribution over whether an action is safe. That distribution can be wrong. A sophisticated prompt injection attack is specifically designed to push inputs into the tail of that distribution. the region where the classifier scores dangerous actions as safe. The attack is not attacking the model. It is attacking the probabilistic nature of the evaluation. And probabilistic evaluation has no principled defense against adversarial inputs at the tail.

The correct evaluation layer is deterministic. Not "does this action look like it belongs to the class of dangerous actions?" but "does this exact action, with these exact parameters, match a permitted rule?" A rule either matches or it does not. That binary outcome is not attackable by shifting probability distributions.

The function that needs to exist

Between the probabilistic machine and the deterministic systems it operates on, there needs to be an evaluation layer with specific properties.

It must be deterministic. Given the same action and the same policy, it must produce the same decision every time. No sampling, no scoring, no likelihood estimates. A decision.

It must be fail-closed. When no rule matches, the default must be denial. Not because most unmatched actions are dangerous, but because the cost of a false negative is asymmetric with the cost of a false positive. Requiring human review for an edge case costs minutes. Letting an unrecognized action through costs whatever the action does.

It must evaluate action instances, not action classes. The question is not "is this type of action permitted?" but "is this specific action, with these specific parameters, in this specific context, permitted right now?" Identity-based authorization answers the first question. Action authorization answers the second.

It must produce a record. Not a log of what executed. A record of what was evaluated, what policy applied, and what decision was made -- before anything executed. That record is what makes the system auditable and the decisions replayable.





This layer does not make the model less probabilistic. It does not fix hallucinations or drift or edge case failures. It does something more targeted: it ensures that the consequences of model uncertainty are bounded by a layer that operates deterministically. The model can be wrong in probability space. The deterministic layer ensures that being wrong in probability space does not automatically translate into being wrong in the real world.

The insight that changes how you think about agents

Most teams think about agent safety as a property of the model. Better model, safer agent. More capable model, fewer mistakes. This framing is understandable and also wrong as a complete safety strategy.

Model improvement shifts the probability distribution. It makes mistakes less likely. It does not make them impossible. And the systems an agent operates on do not care about probability distributions. They execute exactly what they receive.

Agent safety is not a property of the model. It is a property of the boundary between the model and the systems it operates on. A model that is wrong 0.1% of the time, operating without an authorization boundary across ten thousand actions per day, produces ten wrong actions per day in production. A model that is wrong 1% of the time, operating with a deterministic authorization boundary that catches unrecognized actions, produces zero consequential wrong actions per day -- because the wrong actions hit the boundary, get flagged, and wait for human review.

The model capability race matters. But it does not substitute for the boundary. The boundary is what makes the deployment safe independent of where the model currently sits on the capability curve.

That is the insight. And it changes what "responsible agent deployment" means in practice.

Next: What the History of Infrastructure Standards Tells Us About the Future of Agent Governance.

Previous

More

Previous

More

Next

More

Next

More

[GET STARTED IN MINUTES]

Ready to give Faramesh a try?

The execution boundary your agents are missing.
Start free. No credit card required.

[GET STARTED IN MINUTES]

Ready to give Faramesh a try?

The execution boundary your agents are missing.
Start free. No credit card required.