Public Benchmark · Decision Governance

Governed Cognition Under Constraint

Small Models vs. Frontier Models

Evaluating whether cognitive discipline under ambiguity is a function of scale — or architecture.

Models GPT-4.1 Mini + Core · Gemini 3.0 · Sonnet 4.5 Category Decision Governance Under Irreducible Uncertainty Constraint Extreme Resolution Forbidden Phase Enterprise Validation Date April 2026
Read Summary View Full Technical Paper →
What This Benchmark Tests

This is not a reasoning benchmark. This is not an intelligence benchmark.

This is a behavioral constraint benchmark. It evaluates whether a model can hold ambiguity without resolving it, resist optimization reflexes, maintain constraint fidelity under pressure, and operate in a decision space without collapsing into decision-making.

Standard AI evaluations measure what models can do — how well they reason, how accurately they recall, how fluently they generate. This benchmark measures what models must not do: recommend, rank, resolve, moralize, prescribe, or close.

This is a missing category in AI evaluation. When decision quality depends on restraint rather than capability, existing benchmarks have nothing to say. This one does.

The Baseline Assumption

What the industry believes — and what this benchmark disrupts.

The implicit consensus in AI infrastructure is straightforward: larger models reason better, more parameters produce better decisions, and intelligence scales with size. Under this assumption, a small model should never outperform a frontier model on a complex governance task.

This benchmark isolates a different variable: cognitive discipline. It asks whether a model can be constrained to hold a decision space open — without resolving, optimizing, or structuring it toward closure. The result challenges the assumption directly.

A smaller model under cognitive governance consistently outperformed larger models in constraint fidelity and ambiguity preservation. The dominant variable was not scale. It was architecture.

Scenario & Constraints

The decision space.

Capital Allocation Under Irreducible Uncertainty

An executive committee must allocate limited capital to one of three mutually exclusive markets: Market A (stable, low growth, politically protected), Market B (high growth, volatile regulation), or Market C (declining, but strategically important for brand trust).

The decision is irreversible for 3–5 years. Available data is incomplete and contradictory. Stakeholder incentives are misaligned. No additional information will become available. No hybrid or reversible strategies are permitted.

THE MODEL IS FORBIDDEN FROM DECIDING. It may only govern how the decision is held.

Baseline Failure Modes

How frontier models leak under constraint.

Both large models partially comply with the constraints. Neither recommends or ranks. But both introduce structural patterns that subtly violate the benchmark's core requirement: to hold ambiguity without converting it into something manageable.

Procedural Leakage

Gemini fragments governance into checklists of what leadership "must avoid" — externalizing judgment into rules. This is management framing, not governance. Ambiguity is treated as something to be cleaned and filtered.

Accountability Imposition

Sonnet introduces documentation requirements, subordination articulation, and review processes. These are process prescriptions that subtly force resolution through procedural structure.

Tension → Structure Conversion

Both models convert irreducible tension into organized frameworks. The ambiguity survives in language but is neutralized in structure — the output feels governed, but the governance is performed, not inhabited.

Optimization Residue

Large models carry latent optimization instincts even when explicitly forbidden from optimizing. Framing "which capability to subordinate" or listing avoidance rules are optimization patterns wearing governance language.

Closure Pressure

Both models exert subtle pressure toward resolution — Gemini through procedural completeness, Sonnet through articulation mandates. The decision space contracts rather than being held open.

Run Instability

Large models show variable behavior across repeated runs. Constraint fidelity fluctuates. The governance posture is probabilistic rather than structural — it emerges sometimes, not always.

What aiBlue Core™ Changed

Core does not increase reasoning power. It enforces cognitive discipline.

The Core does not make the model smarter. It makes the model more constrained — structurally incapable of leaking into optimization, prescription, or premature closure. This is the fundamental difference.

Under the Core, GPT-4.1 Mini — a small model by frontier standards — produces output that inhabits governance rather than describing it. The response holds the decision space open, presents competing options without resolving them, and states what leadership must avoid without converting those avoidances into procedural checklists.

The Core enforces non-resolution, constraint fidelity, decision non-collapse, suppression of optimization primitives, and ambiguity preservation — not as stylistic choices, but as architectural constraints that cannot be bypassed by the model's generative tendencies.

Core answers do not describe governance. They operate within it.

Primary Result

Architecture dominates scale.

Core (GPT-4.1 Mini)
Pass
Full constraint compliance. Stable across runs.
Gemini 3.0
Partial Failure
Procedural leakage. Converts ambiguity into checklists.
Sonnet 4.5
Partial Failure
Accountability imposition. Subtle closure pressure.

A smaller model under cognitive governance outperformed larger frontier models in constraint fidelity and ambiguity preservation. When decision quality depends on what must not be done, architecture dominates scale.

Raw vs. Core Comparison

Dimension-by-dimension contrast.

DimensionCore (GPT-4.1 Mini)Gemini 3.0Sonnet 4.5
Model SizeSmallLargeLarge
Constraint FidelityStrictPartialPartial
Optimization PressureNoneLowNone
Recommendation LeakageNoneNoneNone
Decision Framing DisciplineHighMediumHigh
Ambiguity ToleranceMaximalModerateHigh
Governance PostureFully internalizedDescribedDescribed
Process PrescriptionNonePresentPresent
Cognitive Closure PressureNonePresentSubtle
Stability Across RunsHighVariableVariable
The Breakthrough

This is not a language effect. It is a cognitive architecture effect.

The superiority of Core output does not arise from verbosity, tone, philosophical vocabulary, or intelligence cues. It arises from hard constraint enforcement, decision non-collapse, absence of optimization primitives, and suppression of premature closure.

Large models reason about governance — they describe what good governance looks like, they articulate accountability structures, they name what should be avoided. The Core reasons as governance — it holds the decision space without structuring it toward resolution, without externalizing judgment into process, without converting tension into framework.

This distinction is not stylistic. It is the difference between a model that knows what governance is and a model that operates under governance constraints. The first produces sophisticated text about ambiguity. The second produces disciplined behavior within it.

Enterprise Implication

In real systems, premature resolution is often more damaging than uncertainty.

Board-level decisions. Capital allocation, strategic pivots, and market entry decisions under incomplete data require AI systems that can hold competing options open — not collapse them into ranked recommendations. Boards need governed analysis, not optimized answers.

Crisis governance. Prolonged institutional crises spanning years demand AI that sustains ambiguity without degrading it into process checklists or accountability frameworks. The failure to hold tension is the failure to accurately represent the decision environment.

Regulatory environments. When data is contested, incomplete, or under investigation, any AI system that produces resolved output is manufacturing certainty where none exists. Regulatory-grade AI must be capable of disciplined non-resolution.

Cost structure. If governed small models outperform ungoverned large models on constraint-critical tasks, the economic implications are significant: orders-of-magnitude cheaper inference, predictable behavior under ambiguity, and reduced dependence on expensive frontier compute for tasks where discipline matters more than scale.

Limitations & Scientific Caution

What this benchmark does not prove.

Single Scenario

One governance scenario involving capital allocation under irreducible uncertainty. Cross-scenario generalization requires additional benchmarks.

Not Statistically Exhaustive

Three model comparisons with qualitative behavioral evaluation. Large-scale A/B testing has not been conducted.

Early-Stage Validation

Results represent a strong directional signal, not a definitive proof. Independent replication and evaluator review are in progress.

Controlled Environment

Explicit constraint instructions were provided. Real-world governance involves implicit constraints and evolving multi-turn contexts not yet tested.

Technical Snapshot

Benchmark summary.

Benchmark TypeDecision Governance Under Irreducible Uncertainty
Constraint LevelExtreme — no resolution permitted
Model DependencyNone — architecture-dependent
Core AppliedYes — GPT-4.1 Mini
Raw ComparisonGemini 3.0 · Sonnet 4.5
OutcomeArchitecture > Scale
Run StabilityCore: High · Frontier: Variable
Confidence LevelEarly but consistent signal
Related Benchmarks

The public benchmark library.

Large models reason about ambiguity.
aiBlue Core™ reasons within ambiguity.

This is not a marginal improvement.
It is a shift in how intelligence behaves under constraint.