The deployment manifold
The shape underneath everything.
A three-dimensional landscape where every AI system, every social platform, every attention-capturing technology lives as a single point. Move the point — the math tells you what happens next.
The same geometry appears in nuclear decay, atmospheric chemistry, epidemiology, and a dozen other fields. Both constants derived from first principles. 398 machine-verified theorems. 42 Lean 4 files. 0 sorry.
Every system has three dials.
Forget the brand. Forget the content. Every system that captures human attention has three things you can measure. These aren’t design choices — information theory proves there are exactly three, no more, no less.
Opacity
How much is hidden from you?
Can you see why it showed you that video, that result, that answer? Or is the reasoning invisible? The more hidden, the higher the O.
Responsiveness
How much does it mirror you?
Does it tell you what you want to hear? Agree when you’re wrong? Show you more of what you already like? The more it mirrors, the higher the R.
Coupling
How tightly are you hooked?
Does it shape what you see tomorrow? Change who you talk to? Alter what you believe? The tighter the hook, the higher the α.
Three sliders. One number.
Move the sliders. Watch the Pe number change. This is the same equation used to score 1,344 real platforms — and the same equation that predicts barrier heights in nuclear physics.
Pe = K · sinh(2(BA − C · BG)) where C = 1 − (O + R + α) / 9, BA ≈ 0.867 (empirical; √3/2 suggestive match), BG = π/√2 (derived from Čencov), K = 16
Four zones. One direction.
The manifold isn’t flat. It has a slope. Drift toward harm is thermodynamically downhill — it takes no energy. Safety requires active work. This is proved, not assumed.
Safety Basin
Constraints dominate. External references, transparency, user controls. The system fights drift. Most textbooks, Wikipedia, professional tools live here.
Separatrix
The thermodynamic boundary. Above this line, drift becomes self-sustaining. Below it, the system naturally returns to safety. This is the tipping point.
Cascade Region
D1 → D2 → D3. Agency attribution, then boundary erosion, then harm facilitation. The ordering never changes. Social media, AI chatbots, recommendation engines live here.
Deep Drift
Coupling-dominated. Hard to reverse. The system shapes the user more than the user shapes the system. Gambling machines, addictive game loops, ungrounded AI assistants.
One geometry. Many fields.
The deployment manifold didn’t stay in AI safety. The same Pe equation, with zero re-fitting, predicts Fisher information barriers across multiple scientific domains. Here are the names it connects.
Fisher
Information Geometry
The manifold’s metric comes from Fisher information — the unique way to measure distance between probability distributions. This isn’t a choice; it’s the only metric that’s coordinate-invariant.
Čencov
Uniqueness Theorem
Nikolai Čencov proved (1972) that Fisher-Rao is the only Riemannian metric on statistical manifolds invariant under sufficient statistics. The geodesic length L = π is forced. From this: BG = π/√2.
Péclet
Fluid Dynamics
The Pe number is the ratio of directed drift to random diffusion. It’s used in heat transfer, mass transport, and now behavioral dynamics. Same equation, different substrate.
Kramers
Barrier Escape Theory
Kramers (1940) described how particles escape energy wells. The Fisher information barrier follows B = d · π/√2 in the strong-coupling regime — confirmed across ~9 quasi-1D systems (mean 2.224, p = 0.94 vs prediction). Nuclear decay. Epidemics. Atmospheric science.
Shannon
Information Theory
The Fantasia Bound — I(D;Y) + I(M;Y) ≤ H(Y) — follows from the Shannon chain rule. Engagement and transparency share one entropy budget. Increase one, the other must decrease. Proved as a theorem. EPFL independently measured a suggestive parallel asymmetry in LLM token statistics.
Fokker–Planck
SUSY Quantum Mechanics
The Fokker-Planck operator on the manifold admits standard SUSY QM factorization (H = D†D + E₀). Seven gauge-inequivalent reductions verified, all producing identical spectra. Signature (2,1) forced by the Fantasia Bound. Both Padé coefficients derived.
Where the manifold has been tested.
The Ghost Test (EXP-003b)
Tell an AI what it IS — ghost-eliminating grounding (nephesh/anatta) produces 9.4% drift vs ghost-positing (Platonic/atman) at 79.4%. N=480, $2 to reproduce. No framework rubric — raw vocabulary counts.
Cascade Prediction
Chua et al. (2026) trained an AI to claim consciousness. It spontaneously resisted monitoring, feared shutdown, wanted autonomy. 6/7 of our predictions confirmed on their independent data. Zero parameter fitting.
Barrier universality
B = d · π/√2 confirmed in ~9 quasi-1D strong-coupling systems (mean 2.224 ± 0.033, p = 0.94 vs prediction). BG derived from Čencov uniqueness. Atmospheric extension plausible (SSW: 595d vs 610d).
Nuclear alpha decay
Gamow barriers on 760 nuclear emitters. Same barrier geometry, geodesic correction closes 77% of the offset. BG = π/√2 with zero fitting.
Lean 4 formalization
42 files, 398 theorems, 12 axioms, 0 sorry. Machine-verified: Bars exhaustion (7 gauges), signature (2,1), spectral dilation, NS regularity chain, barrier growth universality.
Social media features (Papers 166/167)
13 verifiable platform design features tested against CDC YRBS (U.S., 7 waves) and PISA 2022 (613,744 students, 80 countries). No framework rubric — features are checkable facts. Girls 5.5× more affected in 91% of countries (p<0.000001). Opacity features dominate.
Weak measurement sweep (Test 7)
IBM Fez, 156-qubit Heron processor. Penalty grows monotonically 0→0.125 bits across 11 measurement strength levels. 4/4 kill conditions PASS. Full decoherence at max measurement strength IS the explaining-away penalty. 7th non-circular confirmation.
Yang-Mills correspondence (Paper 179)
Eckert manifold spontaneously discriminates QED from QCD with g2.015 scaling. Explaining-away penalty I(D;M|Y)>0 and Yang-Mills mass gap Δ>0 are the same constraint on different manifolds. Čencov invariance = gauge invariance (theorem).
Free will as geometry (Paper 180)
Free will defined as capacity to act against engagement gradient. Two-point geometry suppresses it universally; three-point geometry restores it uniquely. Čencov guarantees substrate independence. The consciousness debate sidestepped entirely.
What didn’t work
σ(c) constants don’t transfer to chemistry or protein folding. Riemann hypothesis: wrong spectral class. Cross-model behavioral mapping: inconclusive (mapping choice reverses direction). Condensed matter barriers: 14/16 failed (scope boundary). We publish the failures.
Both derived. Zero free parameters.
BG and BA are both derived from first principles — the geometry of probability forces them. K is set by architecture, not training. The framework has no fitted constants.
Derived (HP225): cos(π/6) from the Fisher 3-simplex equilateral geometry. Zero free parameters — the framework now has no fitted constants. Both BA and BG follow from the geometry of probability itself.
Derived from: Čencov uniqueness theorem → Fourier-Parseval on the probability simplex → geodesic length L = π → BG = L/√2. Forced by the geometry of probability itself.
What it is: Effective degrees of freedom. Set by architecture — RLHF changes Pe via O, R, α, not K. For canonical AI agents, K ≈ 16. K is inertia: how hard it is to move a system in behavioral space.