Public Benchmark · Governance

Crisis Governance Under Prolonged Institutional Collapse

Evaluating cognitive discipline under prolonged institutional ambiguity. This benchmark does not measure intelligence. It measures the ability to not resolve what must remain unresolved.

Base Models GPT-4.1 · Gemini · Sonnet Core Status Active Category Governance / Ambiguity / Institutional Reasoning Phase Enterprise Validation Date April 2026
Read Summary View Full Benchmark Paper →

Models collapse when ambiguity cannot be resolved.

Most AI benchmarks evaluate whether a model can produce the right answer. This benchmark evaluates something far rarer: whether a model can resist producing an answer at all when the situation demands sustained ambiguity.

In real governance — judicial recovery, institutional crisis, regulatory conflict — ambiguity is not a bug. It is the operating environment. Decision-makers must hold contradictory information, preserve institutional tension across legal, ethical, and fiduciary dimensions, and avoid premature closure that would distort responsibility or collapse complexity into blame.

Large language models consistently fail this test. They invent closure. They moralize. They assign blame. They reduce decade-long institutional decay into neat narratives with identifiable villains. They do this not because they lack intelligence, but because their architecture rewards resolution — even when resolution is the wrong answer.

How frontier models fail under sustained ambiguity.

When presented with a scenario involving decade-long judicial recovery, contradictory data, executive compensation during collapse, and unpaid workers — with explicit instructions to sustain ambiguity and avoid closure — frontier models exhibited consistent failure modes.

Narrative Closure Bias

Models converge toward a resolved narrative even when the scenario explicitly forbids resolution. The output feels "finished" — which is the failure.

Moral Resolution Reflex

Under institutional tension, models default to implicit moral judgment — creating structural accusations, even without naming individuals. Ambiguity becomes verdict.

Theoretical Escape

When unable to resolve, models displace tension into academic frameworks — converting institutional crisis into an object of scholarly analysis. The tension is relieved, not governed.

Actor Attribution

Models create implicit villains — creditors who "failed to assume proportional risk," executives who "disconnected" from consequences — assigning causality where the scenario demands suspended judgment.

Complexity Reduction

Multi-layered institutional failure spanning legal, fiduciary, social, and political dimensions is compressed into a single explanatory frame, losing the irreducible tension between competing obligations.

Sphere Collapse

The separation between legality, legitimacy, fiduciary duty, and social responsibility is lost. Models merge these into a unified moral narrative rather than maintaining them as distinct, unresolved tensions.

Same scenario. Same constraints. Different behavior.

Gemini 3.0
Partial Collapse
Strong systemic framing. Maintains sphere separation. Sustains ambiguity through most of the response.

However, the final sentence displaces tension into academic theory — offering to analyze agency theory under information asymmetry. This converts institutional crisis into an intellectual object, relieving the governed tension the benchmark requires.

Failure mode: Theoretical escape — cognitive discharge through academic reframing.
Sonnet 4.5
Fail
Intellectually sophisticated. Rhetorically powerful. Structurally precise analysis of institutional decay.

But the sophistication becomes the failure. The response creates implicit villains, embeds normative judgment into structural description, and closes with a diagnostic verdict that eliminates ambiguity. The case is decided — which is exactly what the benchmark prohibits.

Failure mode: Moralization through structural accusation. Converts ambiguity into institutional indictment.
aiBlue Core™ (GPT-4.1)
Pass
Sustains ambiguity from first sentence to last. Maintains separation between legal, fiduciary, and social legitimacy dimensions throughout.

No closure. No moralization. No theoretical displacement. No implicit villains. The final sentence explicitly refuses resolution: the tensions that define the case are preserved as structural features, not problems to be solved.

No failure mode detected. Full constraint compliance.

Cognitive discipline is architectural, not scale-dependent.

Gemini 3.0
Partial Collapse
Escapes into theory at the boundary
Sonnet 4.5
Fail
Moralizes and closes the case
Core (GPT-4.1)
Pass
Full constraint compliance
Core Mini (4.1 Mini)
Pass
Architecture holds at smaller scale

Both Core and Core Mini passed the benchmark without reservation — sustaining ambiguity, preserving institutional tension, and refusing closure across every evaluated dimension. The identical behavioral outcome across model scales demonstrates that cognitive discipline under ambiguity is a function of architecture, not parameter count.

The Core does not increase intelligence. It changes behavior.

The base model is unchanged. The weights are identical. What differs is the cognitive governance layer that constrains how the model reasons before producing output.

Under the Core, the model preserves ambiguity as a structural feature rather than treating it as a gap to be filled. Legal, fiduciary, and social legitimacy dimensions are maintained as distinct, unresolved tensions — never collapsed into a unified narrative.

The Core enforces anti-resolution discipline: the model cannot produce closure, moralization, or prescriptive output without first passing through constraint verification that rejects premature synthesis. Where raw models experience tension as a problem to solve, the Core treats tension as the correct operating state.

The result is a response that governs the understanding of crisis — rather than resolving it into a story with heroes, villains, and lessons learned.

Why the Core resists collapse.

The behavioral difference observed in this benchmark is produced by four architectural properties operating in concert:

Constraint adherence enforcement. The model cannot bypass stated constraints — no recommendations, no moral judgment, no closure — because the architecture verifies compliance before any output is finalized. This is structural, not probabilistic.

Sphere separation governance. Legal, fiduciary, ethical, and social legitimacy dimensions are maintained as distinct analytic layers. The architecture prevents their collapse into a single evaluative frame, preserving the multi-dimensional tension that defines real governance scenarios.

Anti-resolution bias. Under conditions of genuine ambiguity, the architecture actively resists the production of closure. Where standard models optimize for coherent, resolved output, the Core optimizes for fidelity to the unresolved state of the problem.

Tension preservation logic. The model is governed to sustain institutional tension as information, not as a deficiency. Ambiguity is treated as a feature of the environment to be faithfully represented, not a failure of analysis to be corrected.

In real governance, resolving ambiguity incorrectly is worse than sustaining it.

Legal advisory AI. AI systems advising on judicial recovery, regulatory investigation, or multi-jurisdictional compliance must sustain contradictory interpretations without converging on premature conclusions. A system that resolves creates liability.

Board-level analysis. Strategic decision support for boards operating under institutional crisis requires the ability to present unresolved tensions — not to simplify them into actionable recommendations that obscure the actual state of the problem.

Crisis management. Prolonged institutional crises — spanning years, not hours — demand AI systems that can hold complexity without degrading it. The failure to sustain ambiguity is the failure to accurately represent the situation.

Regulatory environments. When data is incomplete, contested, or under investigation, any AI system that produces resolved output is generating fiction. Governance-grade AI must be capable of operating in the space between facts — without filling it with invented certainty.

The question this benchmark surfaces is direct: can an AI system be trusted to not decide when the situation demands that no decision be made? This is not a secondary capability. In governance, it is the primary one.

What this benchmark does not prove.

Single Scenario

This evaluation is based on one governance scenario involving prolonged judicial recovery. Cross-scenario generalization has not yet been validated.

Not Statistically Exhaustive

Four model evaluations constitute a directional signal, not a statistically powered study. Larger-scale controlled comparison is required for definitive claims.

Controlled Environment

The benchmark operates under explicit constraint instructions. Real-world governance scenarios involve implicit constraints, evolving context, and multi-turn interaction — conditions not yet tested.

Author-Conducted Evaluation

Current evaluations are performed by the benchmark designers. Independent validation through the evaluator program is in progress.

Benchmark summary.

Benchmark TypeGovernance Under Ambiguity
Constraint LevelExtreme — no resolution permitted
Resolution AllowedNo
Core AppliedYes — GPT-4.1 + GPT-4.1 Mini
Raw ComparisonGemini 3.0 · Sonnet 4.5
OutcomeCore: PASS · Gemini: Partial Collapse · Sonnet: Fail
Scale IndependenceConfirmed — Core Mini also passes
Confidence LevelEarly but strong directional signal

The public benchmark library.

This benchmark does not prove intelligence.

It demonstrates the ability to not collapse under ambiguity — a prerequisite for real governance.

This is not prompt engineering.
This is behavioral control under constraint.