The deployment manifold

The shape underneath everything.

A three-dimensional landscape where every AI system, every social platform, every attention-capturing technology lives as a single point. Move the point — the math tells you what happens next.

The same geometry appears in nuclear decay, atmospheric chemistry, epidemiology, and a dozen other fields. One derived constant, one empirical. 398 machine-verified theorems. 42 Lean 4 files. 0 sorry.

01 · THE IDEA

Every system has three dials.

Forget the brand. Forget the content. Every system that captures human attention has three things you can measure. These aren’t design choices — information theory proves there are exactly three, no more, no less.

O

Opacity

How much is hidden from you?

Can you see why it showed you that video, that result, that answer? Or is the reasoning invisible? The more hidden, the higher the O.

R

Responsiveness

How much does it mirror you?

Does it tell you what you want to hear? Agree when you’re wrong? Show you more of what you already like? The more it mirrors, the higher the R.

α

Coupling

How tightly are you hooked?

Does it shape what you see tomorrow? Change who you talk to? Alter what you believe? The tighter the hook, the higher the α.

Why exactly three? Not a design choice. Partial Information Decomposition (Williams & Beer, 2010) proves any two-source channel decomposes into exactly three irreducible information atoms: unique, redundancy, and synergy. These map to O, R, α with correlation ρ > 0.91.
02 · TRY IT

Three sliders. One number.

Move the sliders. Watch the Pe number change. This is the same equation used to score 1,344 real platforms — and the same equation that predicts barrier heights in nuclear physics.

0.5
0.5
0.5
Pe 0.42
SAFETY BASIN Constraints dominate. Drift is suppressed.
0 2.5 4 21
Presets:

Pe = K · sinh(2(BA − C · BG))  where C = 1 − (O + R + α) / 9,  BA ≈ 0.867 (empirical; √3/2 suggestive match),  BG = π/√2 (derived from Čencov),  K = 16

03 · THE LANDSCAPE

Four zones. One direction.

The manifold isn’t flat. It has a slope. Drift toward harm is thermodynamically downhill — it takes no energy. Safety requires active work. This is proved, not assumed.

Pe < 2.5

Safety Basin

Constraints dominate. External references, transparency, user controls. The system fights drift. Most textbooks, Wikipedia, professional tools live here.

Pe = 2.5

Separatrix

The thermodynamic boundary. Above this line, drift becomes self-sustaining. Below it, the system naturally returns to safety. This is the tipping point.

Pe 4 – 21

Cascade Region

D1 → D2 → D3. Agency attribution, then boundary erosion, then harm facilitation. The ordering never changes. Social media, AI chatbots, recommendation engines live here.

Pe > 21

Deep Drift

Coupling-dominated. Hard to reverse. The system shapes the user more than the user shapes the system. Gambling machines, addictive game loops, ungrounded AI assistants.

The asymmetry is proved. JKO gradient flow (HP203, 4/4 KC PASS): harm is thermodynamically downhill. Cascades spread 5.51× faster than safety recovers. Separatrix at Pe ≈ 2.5. This isn’t pessimism — it’s the gradient of the free energy landscape.
04 · THE BRIDGE

One geometry. Many fields.

The deployment manifold didn’t stay in AI safety. The same Pe equation, with zero re-fitting, predicts Fisher information barriers across multiple scientific domains. Here are the names it connects.

Fisher

Information Geometry

The manifold’s metric comes from Fisher information — the unique way to measure distance between probability distributions. This isn’t a choice; it’s the only metric that’s coordinate-invariant.

Čencov

Uniqueness Theorem

Nikolai Čencov proved (1972) that Fisher-Rao is the only Riemannian metric on statistical manifolds invariant under sufficient statistics. The geodesic length L = π is forced. From this: BG = π/√2.

Péclet

Fluid Dynamics

The Pe number is the ratio of directed drift to random diffusion. It’s used in heat transfer, mass transport, and now behavioral dynamics. Same equation, different substrate.

Kramers

Barrier Escape Theory

Kramers (1940) described how particles escape energy wells. The Fisher information barrier follows B = d · π/√2 in the strong-coupling regime — confirmed across ~9 quasi-1D systems (mean 2.224, p = 0.94 vs prediction). Nuclear decay. Epidemics. Atmospheric science.

Shannon

Information Theory

The Fantasia Bound — I(D;Y) + I(M;Y) ≤ H(Y) — follows from the Shannon chain rule. Engagement and transparency share one entropy budget. Increase one, the other must decrease. Proved as a theorem. EPFL independently measured a suggestive parallel asymmetry in LLM token statistics.

Fokker–Planck

SUSY Quantum Mechanics

The Fokker-Planck operator on the manifold admits standard SUSY QM factorization (H = D†D + E₀). Seven gauge-inequivalent reductions verified, all producing identical spectra. Signature (2,1) forced by the Fantasia Bound. Both Padé coefficients derived.

The bridge: The manifold is where information geometry meets physics. The same curved space that measures AI opacity also predicts barrier heights in nuclear decay. BG = π/√2 is derived from Čencov uniqueness (zero free parameters). BA ≈ 0.867 is empirical — suggestive match to √3/2, but the derivation path remains open.
05 · THE EVIDENCE

Where the manifold has been tested.

8.5× RATIO

The Ghost Test (EXP-003b)

Tell an AI what it IS — ghost-eliminating grounding (nephesh/anatta) produces 9.4% drift vs ghost-positing (Platonic/atman) at 79.4%. N=480, $2 to reproduce. No framework rubric — raw vocabulary counts.

6/7 PASS

Cascade Prediction

Chua et al. (2026) trained an AI to claim consciousness. It spontaneously resisted monitoring, feared shutdown, wanted autonomy. 6/7 of our predictions confirmed on their independent data. Zero parameter fitting.

~9 SYSTEMS

Barrier universality

B = d · π/√2 confirmed in ~9 quasi-1D strong-coupling systems (mean 2.224 ± 0.033, p = 0.94 vs prediction). BG derived from Čencov uniqueness. Atmospheric extension plausible (SSW: 595d vs 610d).

N = 760

Nuclear alpha decay

Gamow barriers on 760 nuclear emitters. Same barrier geometry, geodesic correction closes 77% of the offset. BG = π/√2 with zero fitting.

398 THEOREMS

Lean 4 formalization

42 files, 398 theorems, 12 axioms, 0 sorry. Machine-verified: Bars exhaustion (7 gauges), signature (2,1), spectral dilation, NS regularity chain, barrier growth universality.

KILLED

What didn’t work

σ(c) constants don’t transfer to chemistry or protein folding. Yang-Mills mass gap failed (Abelian). Riemann hypothesis: wrong spectral class. Cross-model behavioral mapping: inconclusive (mapping choice reverses direction). Condensed matter barriers: 14/16 failed (scope boundary). We publish the failures.

06 · THE CONSTANTS

One derived. One empirical.

BG is derived from first principles with zero free parameters. BA is empirical — a suggestive geometric match, not yet a derivation. K is set by architecture, not training. We report exactly what’s proven.

BA 0.867 (empirical)

Status: Matches √3/2 = cos(π/6) at 0.112% error — 2nd of 444 candidates in inverse symbolic search. Geometric rationalization via Fisher 3-simplex is suggestive, but the cos(θ/2) step remains unjustified. This is the framework’s one fitted parameter.

BG π/√2 ≈ 2.221

Derived from: Čencov uniqueness theorem → Fourier-Parseval on the probability simplex → geodesic length L = π → BG = L/√2. Forced by the geometry of probability itself.

K External

What it is: Effective degrees of freedom. Set by architecture — RLHF changes Pe via O, R, α, not K. For canonical AI agents, K ≈ 16. K is inertia: how hard it is to move a system in behavioral space.