Core Thesis
Systems differ not by apparent complexity, but by consequence geometry—how actions map to futures.
A system is deep if: Small local actions sharply collapse the future state space
A system is shallow if: Local errors preserve most futures and can be averaged away
Intelligence (minimally defined as optimisation over futures) succeeds where systems are diagonalisable.
History breaks only where diagonalisation fails.
A Note on Language
This essay uses mathematical terminology (eigenvectors, diagonalisation, basis change) not as metaphor but as precise structural description. If you’re unfamiliar with linear algebra:
- Eigenbasis = the fundamental coordinates/patterns that explain how a system behaves
- Diagonalisable = can be understood as a sum of independent, stable patterns
- Basis change = when the fundamental categories you use to describe reality stop working
Think of it this way: if you’re navigating a city, the eigenbasis is “streets and buildings.” A basis change would be if the city suddenly operated like a 3D network (flying cars) where “street addresses” become meaningless—you’d need entirely new coordinates.
1. Diagonalisation as the Structural Test
What diagonalisation means here (non-metaphorical)
A system is diagonalisable if:
- Behaviour can be decomposed into independent modes
- Global dynamics ≈ weighted sum of dominant eigenvectors
- Noise averages out
- Optimisation converges to stable attractors
- Repetition reinforces structure
Canonical cases:
- PageRank on graphs
- Spectral methods on networks
- Normal modes in physics
- Central limit behaviour in statistics
Key rule: If a system is diagonalisable, optimisation eliminates surprise.
2. PageRank as the Prototype
PageRank works because:
- The web graph has dominant eigenmodes
- Repeated reinforcement concentrates visibility
- Peripheral variation decays
Outcomes:
- Centrality becomes a fixed point
- Power-law hierarchies emerge
- Marginal deviation does not alter ranking
This is not a web-specific quirk. It is a generic property of smooth systems with low consequence curvature.
3. Apparent Complexity vs Structural Rank
Systems that feel complex but are low-rank
Music, language, style, culture, fashion, taste
They exhibit:
- High surface variation
- Real skill gradients
- Local sensitivity
- Rich phenomenology
But structurally:
- Errors smear, not cascade
- Futures remain open
- Recovery is cheap
- Averaging improves outcomes
- Dominant eigenmodes exist
These systems are wide but shallow. They feel deep precisely because they forgive error.
4. Systems That Resist Diagonalisation
Some systems are hostile to smoothing:
- Mathematics
- Strategy games
- Engineering
- Legal commitments
- War
- Infrastructure
Properties:
- Small errors annihilate futures
- Local mistakes propagate globally
- No averaging principle
- No stable eigenbasis
But the brittleness has different structural sources:
Mathematics: Chain dependencies with no redundancy (one broken link invalidates the entire proof)
Engineering: Hard physical constraints (10% structural weakness ≠ 10% worse performance, it means collapse)
War: Adversarial optimization (errors get exploited rather than averaged)
Intelligence struggles here not because of scale or complexity, but because approximation destroys validity.
5. History as a Mostly Diagonalisable Object
This motivates psychohistory (non-sci-fi):
At large N:
- Individual actions decorrelate
- Aggregate behaviour stabilises
- Noise averages out
History acquires:
- Eigenmodes (stable patterns)
- Long trends
- Statistical regularity
Consequences:
- Empires rise and fall predictably (resource extraction → overextension → collapse)
- Economic cycles recur (boom → speculation → bust → recovery)
- Cultural convergence dominates (writing, cities, metallurgy emerge independently)
- “Great men” rarely matter structurally
Empirical examples:
- The Bronze Age Collapse (~1200 BCE): Multiple civilizations fell simultaneously through similar dynamics (climate stress + systems interdependence), despite minimal contact
- Agricultural revolution: Emerged independently in at least 7 different regions within a few thousand years
- State formation: Similar institutional patterns emerge across unconnected societies (taxation, bureaucracy, writing systems)
The historiographical caveat:
This is not claiming history is deterministic—contingency matters immensely at human timescales. Rather, at sufficient scale and aggregation, patterns emerge that individuals cannot override. Rome didn’t have to fall in 476 CE, but an empire with that structure, facing those resource constraints, was statistically likely to fragment within some window.
The strongest counterargument comes from “long-tail” historical events—rare occurrences (Genghis Khan, the Black Death, Columbian exchange) that do reshape trajectories. But note: these are often either exogenous shocks (plague, climate) or endogenous Mules (see Section 8), not refutations of the framework.
History is mostly diagonalisable—which is precisely why true Mules matter.
6. Why the “Great Man” Mule Fails (Usually)
The classic Mule (singular individual) is wrong in most contexts:
Remove the individual → The future class usually survives. Another actor occupies the role.
Examples of structural replaceability:
- Remove Napoleon → Another general rides French Revolutionary energy (the structural forces: mass conscription, revolutionary ideology, European imbalance of power)
- Remove Steve Jobs → Computing revolution continues (GUI, personal computing, mobile were structural inevitabilities)
- Remove Einstein → Relativity emerges (Poincaré, Lorentz were converging on the same mathematics)
Individuals ride gradients. They do not create new consequence geometry.
When individuals DO matter:
Not when they’re personally exceptional, but when they catalyze coordination at critical thresholds.
The role is replaceable in principle but may not be filled in practice because:
- Coordination windows are narrow
- Multiple simultaneous conditions must align
- Historical accidents determine who occupies catalyst positions
Example: Lenin in 1917
- Remove Lenin → Russian Revolution might still occur (Tsarist collapse was structural)
- But Bolshevik victory was contingent on specific coordination at specific moments
- Lenin didn’t create revolutionary conditions, but he may have determined which equilibrium Russia fell into
The framework doesn’t deny individual agency—it specifies when it matters: at coordination thresholds near unstable equilibria. Most of history isn’t near such thresholds.
A real Mule must:
- Reassign which actions have irreversible effects
- Alter the dimensionality of the state space
That cannot be an individual property—but individuals can sometimes trigger basis changes that would not otherwise occur (or would occur much later/differently).
7. Definition of a True Mule
(The term “Mule” comes from Asimov’s Foundation series, where a single mutant individual disrupts the predictions of psychohistory—the mathematical sociology that makes civilizational outcomes predictable. Here we use it more precisely to mean any event that breaks the predictive structure itself.)
A Mule is an event or capability that destroys the existing eigenbasis of history.
Operationally:
- Old modes stop spanning the future
- Prior optimisation becomes incoherent
- The system is no longer diagonalisable in its old coordinates
8. Two Classes of Mules
A. Exogenous Mules
- Originate outside the system
- Invisible to internal optimisation
- Maximal consequence curvature
- Reset the game entirely
Examples: Asteroid impacts, supervolcanoes, ice ages
These redefine the fitness function itself.
B. Endogenous Mules (the critical case)
Properties:
- Visible in outline
- Predictable in principle
- Pathologically hard to reach
- Singularities in capability space
Shared features:
- Long flat fitness valleys
- Weak or negative intermediate payoff
- High coordination thresholds
- Sudden payoff activation
- Post-threshold system reorganisation
These are not surprises—they are tunnelling events.
9. The Eye as the Canonical Endogenous Mule
Structurally important because:
Vision is obviously useful. End state is imaginable. “Tech tree” can be sketched.
But:
- Early stages confer minimal advantage
- Costs precede benefits
- Selection gradients are weak
- Most evolutionary paths fail
The basis change was not “seeing”—it was transforming the environment itself.
Before vision:
- Distance protected you from predators
- Concealment was reliable
- Most information was local (touch, chemistry)
- The fitness landscape was one shape
After vision:
- Distance no longer protects
- Concealment becomes an arms race
- Information becomes non-local
- The entire ecology reorganises around information warfare
This is not just adding a capability—it’s redefining what capabilities mean.
Predation, camouflage, signaling, mate selection—every optimization strategy had to be rebuilt. The eigenbasis of “survival” changed coordinates.
Why tunnelling succeeds at all:
Not all lineages cross this barrier. The eye evolved independently ~40 times, but failed in most branches.
Tunnelling succeeds through:
- Population size (more parallel paths explored)
- Neutral drift (wandering across flat landscapes)
- Exaptation (intermediate forms serve other functions—light sensitivity aids circadian rhythm before it enables vision)
- Environmental context (certain niches make the valley shorter)
The question is not whether tunnelling is possible, but what conditions make it probable within historical time.
10. Why Tech Trees Are Misleading
Tech trees get one thing right: Capabilities, not agents, shape destiny
They get one thing wrong: They make the future legible in advance
Tech trees:
- Enumerate outcomes
- Hide reachability
- Suppress epistemic shock
- Eliminate true singularities
A Mule that can be named in advance is already domesticated.
11. Civilization’s Hidden Limit
Civilization (the game) already is a combinatorial technology game. That is not what’s missing.
What Civilization does correctly
- Nonlinear prerequisites
- Cross-tree synergies
- Contextual acceleration
- Soft path dependence
Where Civilization stops short
- All abstractions are enumerable
- The representational space is fixed
- Categories never mutate
Civ allows: Combinatorial unlocks
Civ forbids: Combinatorial abstraction
12. Linear Algebra Translation (Precise)
Civilization explores a fixed vector space:
- New basis vectors are unlocked
- Old ones strengthened or weakened
- The basis itself never changes
In simpler terms: Imagine describing your location. In a 2D city, you use two coordinates (North-South, East-West). Adding a subway system adds a new basis vector (which line you’re on), but you’re still using the same type of description—discrete locations connected by routes.
A basis change would be like switching to a description where “location” stops meaning “a fixed point” at all—perhaps everyone is constantly moving, and you describe positions relative to other moving objects. The old coordinate system (street addresses) can’t even express the new reality.
Civilization (the game):
A Mule is not:
- A deep node (unlocking “Nuclear Fission” makes you powerful)
- A hard-to-reach tech (requires many prerequisites)
- A powerful unlock (gives you strategic advantage)
A Mule is: A basis change, not a basis expansion.
What this would actually look like:
A real Mule in Civ terms would make:
- “Production per turn” stop being meaningful (perhaps everything is now continuous-time)
- “Territory control” become incoherent (perhaps power is now network-based, not geographical)
- “Military units” cease to be the right abstraction (perhaps conflict is now informational/economic)
The UI couldn’t display it. The balance couldn’t accommodate it. The gameplay would break.
This is why Civilization never mutates representation—and why it can’t model true historical discontinuities.
13. What a Real Mule Would Do (Structurally)
In Civ-like terms, a true Mule would cause:
- Resources to change interpretation
- Units to stop being the right abstraction
- Borders to lose explanatory power
- Cities to become administrative nodes
- Power to migrate to new representations
These are representation changes—not buffs, not synergies, not unlocks.
Civilization never mutates representation—hence no true Mules.
14. Why This Is Not a Design Failure
Players require stable abstractions. UI depends on conserved categories. Balance assumes legibility. Learnability forbids basis collapse.
Therefore: Civilization models history after legibility, not history as lived.
This is necessary domestication.
15. The False Mule (Negative Control)
Definition
A false Mule appears to threaten the system but ultimately reinforces the same eigenbasis.
Properties:
- Highly narrativised
- Ideologically charged
- Rapid adoption
- Strong believers and opposition
But structurally:
- No basis change
- No reassignment of irreversible consequence
- Existing optimisation strategies still work
- Institutions adapt without mutation
Canonical False Mule: Cryptocurrency
Structural analysis:
- Money remains scalar and fungible
- Value remains denominated against legacy systems
- States retain violence, law, taxation
- Centralisation re-emerges
- Power-law hierarchies persist
Markets absorb it. Disruption without re-coordination.
Diagnostic Test
Does this force dominant actors to abandon their optimisation strategies?
If they can adapt, capture, regulate, or incorporate it → not a Mule.
A real Mule makes optimisation fail, not adjust.
16. The Printing Press (Calibration Example)
Was the printing press a Mule?
Yes, but a slow one.
Initially:
- Fit existing abstractions (books were still books, just cheaper)
- Markets absorbed it (scribes → typesetters)
- Power structures adapted (licensing, censorship)
But over centuries:
- Made “information scarcity” incoherent as an organizing principle
- Enabled coordination without institutional control
- The eigenbasis of “Church mediates truth” stopped spanning the state space
The Reformation happened because:
- Printing + vernacular Bibles = new coordination modes
- Individual conscience became a valid abstraction
- National churches emerged as alternatives
Why was the basis change so gradual?
The printing press didn’t instantly collapse the old eigenbasis because:
- Literacy rates remained low (most people couldn’t read for generations)
- Institutional power had slack (multiple levers: military, economic, social)
- The technology needed complementary changes (paper production, literacy education, vernacular translation)
But as these accumulated, the rate of basis change accelerated—Protestant Reformation (1517) came ~70 years after Gutenberg (~1440), a rapid collapse once critical mass was reached.
This suggests Mules exist on a spectrum:
- Instant Mules: Nuclear weapons (eigenbasis collapse in years) Why rapid: No intermediate adaptation possible—either you have them or you don’t, game theory completely changes
- Fast Mules: Industrialization (decades) Why rapid: Factory system incompatible with feudal labor relations, forced rapid restructuring
- Slow Mules: Printing press (centuries) Why gradual: Old institutions had slack, complementary technologies needed time, network effects required scale
- False Mules: Cryptocurrency (eigenbasis intact after decades) Why false: Existing power structures can adapt without changing fundamental coordinates
The rate of eigenbasis collapse determines the violence of historical disruption. Fast collapses (industrialization, nuclear weapons) produce revolutionary upheaval. Slow collapses (printing) produce gradual institutional evolution punctuated by crisis moments.
17. Why False Mules Are Inevitable
Optimisation pressure is high. Systems seek release. Innovation clusters near boundaries. Boundary crossing is punished.
So systems generate disruptions that feel radical but remain representationally safe.
False Mules are structural decoys, not conspiracies.
18. Candidate Endogenous Mules (Future)
These are not predictions, only latent singularities.
Mule Candidate 1: Programmable Sovereignty
- Power detaches from territory
- Law becomes protocol-bound
- Citizenship ceases to be scalar
Breaks: Nation-state eigenbasis, border-based optimisation
Mule Candidate 2: Cognitive Labour Collapse
- Thought ceases to be the unit of value
- Skill gradients flatten
- Attribution dissolves
Breaks: Career optimisation, education → productivity mapping
Mule Candidate 3: Ungovernable Energy Abundance
- Energy becomes locally abundant
- Chokepoints dissolve
- Capture fails
Breaks: Capital accumulation, infrastructure leverage, scale dominance
All three are:
- Visible in outline
- Unrewarded in transition
- Structurally hostile to optimisation
19. Why Optimisation Eliminates Its Own Escape Routes
The processes that optimise a system within a regime necessarily destroy that system’s capacity to exit the regime.
This is not a contingent failure. It is a consequence of diagonalisation itself.
Optimisation strengthens eigenbases
Optimisation requires:
- Stable objective functions
- Conserved abstractions
- Repeatable success criteria
- Reinforcement through iteration
Under these conditions:
- Dominant eigenmodes strengthen
- Variance collapses
- Peripheral representations decay
- Noise is actively suppressed
- The system becomes increasingly diagonalisable
This is not accidental. It is what optimisation is.
As optimisation improves, the system becomes more predictable, more efficient, and more legible—and therefore less capable of representational change.
Exploration is structurally opposed to optimisation
Exploration requires:
- Illegible or undefined payoffs
- Persistence without justification
- Tolerance of systematic failure
- Preservation of unused degrees of freedom
- Acceptance of non-convergent behaviour
These properties are incompatible with mature optimisation.
Optimisation and exploration are antagonistic at the level of representation, not merely trade-offs along a spectrum.
20. How Endogenous Mules Are Actually Crossed
Why in-regime optimisation cannot reach Mules
An endogenous Mule lies behind a region with these properties:
- No reliable gradient points toward it
- Intermediate steps are unrewarded or punished
- Coordination payoffs are undefined
- Success cannot be distinguished from noise in advance
Any system that demands efficiency, penalises deviation, requires justification at each step, and eliminates redundancy will systematically avoid these trajectories.
This is not a failure of intelligence, foresight, or imagination. It is a structural consequence of in-regime optimisation.
Meta-optimisation with orthogonal objectives
Endogenous Mules are crossed only by optimisation processes whose objectives do not bottleneck through the current eigenbasis.
Examples:
Evolution optimises for population persistence, not individual fitness
- Uses parallelism (many lineages explore simultaneously)
- Uses neutrality (drift across flat landscapes)
- Uses exaptation (intermediate steps serve other functions)
Science optimises for explanatory compression, not immediate utility
- Tenure protects non-optimization
- Paradigm shifts occur when anomalies accumulate
- Revolutionary science is not deliberate—it’s responsive to eigenbasis breakdown
Markets (at their most disruptive) optimise for option value, not expected return
- Bubbles fund exploration that “rational” allocation wouldn’t
- VC tolerates 90% failure for 10% breakthrough
- Bankruptcy separates exploration cost from system survival
Critical insight: These are still optimisation processes, but their objective functions are orthogonal to the dominant representation. Variance is preserved as a structural feature, not a tolerated inefficiency.
Endogenous Mules are crossed despite in-regime optimisation, not because of it.
21. The Maturity Trap (Formal Statement)
As a system matures, it converts representational flexibility into efficiency. This conversion is irreversible under continued optimisation.
Consequences:
- Mature systems ossify
- Dominant abstractions become self-reinforcing
- Alternative representations are systematically eliminated
- Transformative change becomes statistically invisible
The system is not stagnant by accident. It is too well optimised to escape its own coordinates.
22. Intelligence and Regime Boundaries
This yields a sharp and uncomfortable conclusion:
Intelligence, defined as optimisation over a given future space, cannot navigate basis changes. It can only survive them once they occur.
Corollaries:
- Arbitrarily powerful intelligence remains regime-bound
- No amount of foresight allows deliberate targeting of endogenous Mules
- Transformative change is necessarily: accidental, wasteful, partially blind
- Steering is possible only at the meta-level: preserving variance, not selecting outcomes
23. Detecting Eigenbasis Breakdown
You cannot detect Mules directly, but you can detect when your current eigenbasis is becoming incoherent.
Observable signatures of approaching boundaries:
1. Anomaly accumulation without resolution
- Repeated failures that don’t respond to increased optimisation
- Problems that get worse as you apply more resources
- Example: Pre-revolutionary France—more taxation → less revenue
2. Coordination breakdown despite aligned incentives
- Actors with identical goals cannot agree on strategies
- Every proposed solution creates new problems
- Example: Late-stage USSR—every reform contradicted others
3. Success/failure become illegible
- Cannot distinguish good performance from lucky noise
- Winners cannot explain why they won
- Example: Venture capital pre-2000 bubble
4. Rapid capability discontinuities
- Small changes in inputs → disproportionate changes in outputs
- System sensitivity increases dramatically
- Example: Nuclear weapons—gap between “nearly working” and “working” was months
5. Meta-model breakdown
- Models of why your models work stop working
- Paradigm defense becomes more common than paradigm use
- Example: Ptolemaic astronomy—increasingly elaborate epicycles
The operational test
In a diagonalisable regime:
- Anomalies get resolved by better optimisation
- Coordination failures indicate misaligned incentives
- Success is attributable and reproducible
- Capabilities scale predictably
- Meta-models strengthen over time
Near a Mule:
- Anomalies persist despite optimisation
- Coordination fails despite aligned incentives
- Success is contextual and illegible
- Capabilities jump discontinuously
- Meta-models become defensive
Detection criterion: Are your problems getting more soluble or less soluble as you apply more intelligence?
If more soluble → optimise harder
If less soluble → you’re approaching a boundary, preserve optionality
24. The Conditional Prescription
“Preserve optionality” is not a universal prescription. It is a conditional prescription triggered by detectable symptoms of eigenbasis breakdown.
Normal operation (inside regime):
- Monitor for eigenbasis breakdown signatures
- If problems become more soluble with optimisation → optimise aggressively
- Maintain minimal optionality insurance (hedge against undetected boundaries)
Approaching a boundary:
- When anomalies accumulate without resolution → reduce optimisation intensity
- Shift from exploitation to exploration
- Increase optionality preservation (even if expensive)
- Avoid premature convergence on any single model
At the boundary:
- You cannot predict which direction to go
- You cannot optimise your way through
- All you can do is: survive the crossing, maintain representational flexibility, recognise new eigenmodes after they emerge
After crossing:
- New eigenbasis becomes apparent in hindsight
- Resume optimisation in new coordinates
- Gradually reduce optionality overhead as new regime stabilises
The key behaviours near boundaries:
- Maintaining heterogeneous models
- Tolerating inefficiency
- Allowing apparently irrational persistence
- Avoiding premature convergence
These behaviours appear wasteful inside a regime. They are the only behaviours that survive regime change.
28. Personal and Organizational Implications
This framework isn’t just macro-historical—it applies at every scale.
For individuals:
In diagonalisable domains (most of life):
- Optimize hard
- Learn from feedback
- Build on expertise
- Errors are recoverable
Examples: Career development in stable industries, skill acquisition in established fields, financial planning in normal markets
Near personal Mules:
- Career transitions where old skills become irrelevant
- Relationship dynamics where communication patterns stop working
- Health crises where recovery isn’t “getting back to normal”
Signature: You’re working harder but getting worse results. More effort doesn’t resolve the problem—it intensifies it.
Response: Stop optimizing in the old coordinates. Preserve flexibility. Experiment with different frames. Accept that past success doesn’t predict future success.
For organizations:
In mature markets (diagonalisable):
- Process optimization works
- Best practices compound
- Metrics guide decisions
- Efficiency drives success
Approaching market Mules:
- Kodak and digital photography (optimization in film chemistry became irrelevant)
- Blockbuster and streaming (optimization of retail locations became irrelevant)
- Traditional media and social platforms (optimization of editorial curation became irrelevant)
Diagnostic: Your competitors aren’t playing your game. Your key metrics stop correlating with success. Industry veterans can’t explain why new entrants win.
Response (Christensen’s insight refined): The issue isn’t “disruption from below”—it’s that the basis itself is changing. You can’t defend against this by being better at the old game. You need parallel exploration in new coordinate systems.
For small-scale systems:
When to optimize:
- Stable relationships (communication patterns converge)
- Established routines (feedback loops are clear)
- Known domains (expertise compounds)
When to preserve optionality:
- New relationships (don’t know what matters yet)
- Life transitions (old patterns may not transfer)
- Novel situations (success criteria unclear)
The practical heuristic:
Ask: “If I keep doing what’s working, will I get closer to my goal?”
- Yes → You’re in a diagonalisable regime, optimize
- No, but I can see the problem → Adjust strategy, still diagonalisable
- No, and I can’t tell why → Possibly near a basis change, preserve flexibility
The “premature optimization” error:
Attempting to optimize before you know the eigenbasis is a form of premature convergence. This is why:
- Startups that “pivot” often succeed (they’re exploring the basis)
- Startups that “execute perfectly” on wrong ideas fail (they optimized before finding the eigenbasis)
- Scientific fields progress through paradigm shifts, not just accumulation
The skill is recognizing which regime you’re in—and most errors come from applying optimization when you should be exploring, or vice versa.
Using the detection mechanism on present conditions:
Evidence of eigenbasis coherence (optimise hard):
- Tech still scales predictably (Moore’s law variants)
- Markets still efficiently allocate capital in most domains
- Coordination still works for aligned actors in many contexts
Evidence of eigenbasis breakdown (preserve optionality):
- AI capabilities: Rapid, discontinuous jumps (GPT-2 → GPT-3 → GPT-4)
- Coordination: Increasing difficulty despite aligned incentives (climate, biosecurity, AI governance)
- Success legibility: Decreasing (why do some companies/countries/policies succeed where others fail?)
- Meta-models: Increasingly defensive (economic theories, political ideologies all under strain)
Diagnosis: We are likely approaching a boundary, but not yet at it.
Implication: This is the regime where optionality preservation becomes high-value, even at significant efficiency cost.
Which means:
- Institutional diversity matters more than institutional optimisation
- Distributed experimentation matters more than coordinated strategy
- Maintaining contradictory models matters more than achieving consensus
29. Current Trajectory Assessment
Iain M. Banks clearly intuited that sufficiently advanced intelligence smooths history. His Culture novels are saturated with this insight: overwhelming optimisation power dampens conflict, absorbs shocks, and renders individual human agency largely irrelevant.
What Banks never specifies is the failure mode.
His “Outside Context Problems” function as narrative shocks, but they are almost always exogenous and ultimately legible to superior intelligence. They do not destroy the Culture’s abstractions, invalidate its optimisation strategies, or force a change of representational basis.
The Minds may lose tactically; they never lose the model.
In the terms used here: the Culture has enemies, but it never has a Mule.
Banks describes history after diagonalisation has succeeded. He does not characterise the structural conditions under which diagonalisation must fail.
That omission is not a literary flaw—but it marks the boundary between intuition and theory.
32. Visual Guides to Key Concepts
Diagonalization vs Non-Diagonalizable Systems
DIAGONALIZABLE SYSTEM (e.g., Music, Language)
Error Input: ●──────────────────────────────────────▶
│ Small mistakes
│
Future Space: │ ████████████████████████████ ← Most futures preserved
│ ████████████████████████████
│ ███████●█████████████████████ ← Error absorbed
│ ████████████████████████████
└────────────────────────────────────────▶
Properties:
- Errors "smear" across future space
- Dominant eigenmodes (stable patterns) remain
- Averaging improves outcomes
- System forgives exploration
NON-DIAGONALIZABLE SYSTEM (e.g., Mathematics, Engineering)
Error Input: ●──────────────────────────────────────▶
│ Small mistakes
│
Future Space: │ ████████████████████████████
│ ████████████████████████████
│ ███●───────────────────────── ← Future collapses
│ ───────────────────────────── (Invalid region)
└────────────────────────────────────────▶
Properties:
- Errors cascade and eliminate futures
- No stable eigenbasis
- Approximation destroys validity
- System punishes deviation
Basis Change vs Basis Expansion
BASIS EXPANSION (Civilization-style tech trees)
Before: After:
Dimension 1 ──▶ Dimension 1 ──▶
Dimension 2 ──▶ Dimension 2 ──▶
Dimension 3 ──▶ (NEW - unlocked)
State space: [x, y] → [x, y, z]
Old coordinates still work, just more powerful
BASIS CHANGE (True Mule)
Before: After:
North-South ──▶ Momentum ──▶
East-West ──▶ Phase ──▶
State space: [position] → [wavefunction]
Old coordinates become incoherent
The Eye Evolution: Fitness Landscape
FITNESS LANDSCAPE (simplified 2D projection)
Fitness
↑
│ ╱▔▔▔▔▔▔▔╲
│ ╱ ╲ ← Vision
│ ╱ ╲ (high fitness)
│ ▁ ╱ ╲
│ ╱ ╲ ╱ ╲
│ ╱ ╲ ← Chemosensitivity│ │
│ ╱ ╲ (local peak) │ │
│ ╱ ╲ │ │
│___╱_________╲_________________│_________________│_______
│ ╲________________╱ ← Flat valley │
│ (no fitness (costly, no │
│ gradient) intermediate │
│ benefit) │
└──────────────────────────────────────────────────────▶
Complexity
BEFORE VISION:
- Distance = protection
- Environment: local information dominant
- Fitness landscape: one geometry
AFTER VISION:
- Distance ≠ protection (information is non-local)
- Environment: transformed into information warfare
- Fitness landscape: entirely new geometry
- All optimization strategies must be rebuilt
This is not "adding a capability"—it's changing what capabilities mean.
Detecting Eigenbasis Breakdown
STABLE REGIME INDICATORS BOUNDARY PROXIMITY INDICATORS
Anomalies ──▶ Resolve Anomalies ──▶ Accumulate
with optimization despite optimization
Coordination Success Coordination Failure
●────●────● ● ● ●
│ │ │ │ ╱ │ ╲ │
●────●────● ● ● ●
(aligned actors (aligned goals,
achieve goals) can't coordinate)
Success Metrics Success Metrics
Input ──▶ Output Input ──?──▶ Output
(predictable (illegible
attribution) causation)
Meta-Models Meta-Models
┌──────────┐ ┌──────────┐
│ Theory │──▶ Stronger │ Theory │──▶ Defensive
│ explains │ │ can't │
└──────────┘ │ explain │
└──────────┘
DECISION RULE:
Are problems becoming MORE or LESS soluble with optimization?
├─ More soluble → Optimize harder (stable regime)
└─ Less soluble → Preserve optionality (approaching boundary)
Mule Spectrum: Rate of Eigenbasis Collapse
INSTANT MULE (years)
Nuclear Weapons
│
├── Old eigenbasis: "War = large armies + territory"
├── Instant collapse: "War = mutually assured destruction"
├── No intermediate adaptation possible
└── Complete re-coordination required
Time: ~5 years (1945-1950)
FAST MULE (decades)
Industrialization
│
├── Old eigenbasis: "Production = skilled craft labor"
├── Gradual collapse: "Production = factory system"
├── Institutions forced to adapt rapidly
└── Social upheaval, but not instant
Time: ~30-50 years (1780s-1830s)
SLOW MULE (centuries)
Printing Press
│
├── Old eigenbasis: "Information = scarce, Church-mediated"
├── Very gradual collapse: "Information = abundant, distributed"
├── Institutions had slack to adapt incrementally
└── Crisis moments (Reformation) punctuate slow change
Time: ~200 years (1450-1650)
FALSE MULE (no collapse)
Cryptocurrency
│
├── Appears to threaten: "Money = state-issued currency"
├── Actually reinforces: Same eigenbasis persists
├── Markets absorb without basis change
└── Disruption without re-coordination
Time: 15+ years, eigenbasis intact
RATE DETERMINANT: How much can the old eigenbasis accommodate
before fundamental categories stop working?
Smooth systems:
- Diagonalisable
- Eigenmodes dominate
- Optimisation succeeds
- History feels inevitable
Deep systems:
- Non-diagonalisable
- High consequence curvature
- Optimisation fails locally
True historical breaks:
- Occur when abstraction mutates
- Destroy the existing eigenbasis
- Create new axes of optimisation
33. Conclusion
Intelligence does not create depth.
It eliminates depth wherever it can.
History is smooth wherever optimisation succeeds—and discontinuous only where the geometry of consequence itself refuses to be flattened.
Optimisation strengthens eigenbases. Therefore, systems that optimise successfully necessarily reduce their capacity for basis change.
Historical discontinuities occur when consequence geometry forces basis change despite optimisation resistance.
This is the inversion that makes intelligence both powerful and bounded: it flattens landscapes until it encounters geometry that cannot be flattened—and there, necessarily, it breaks.

