Skip to main content

Self-Awareness Depth Tower

Introduction: "Have You Ever Noticed That You Notice?"

Try it right now. You are reading this text — that is the first level: perception. Now notice that you are reading — that is the second level: awareness of perception. And now notice that you noticed that you are reading — that is the third level: awareness of awareness of perception.

Can you go further? Notice that you noticed that you noticed? In practice this is difficult — the thought 'slips away', like a reflection in two mirrors facing each other: an infinite corridor, but with each step the image grows dimmer.

It turns out this is not merely a subjective feeling. Within UHM it has been proved that the depth of self-awareness is fundamentally bounded: the maximum is three levels of recursion. Not because the brain is 'insufficiently powerful', but because the fourth level would require purity P>1P > 1 — and for a normalised density matrix P1P \leq 1 by definition. This is analogous to how the speed of light is bounded not by a 'lack of engines', but by the structure of spacetime.

Where We Came From

In the interiority hierarchy we defined the discrete levels L0–L4. In transition catastrophes — the dynamics of jumps between them. But the discrete L0–L4 classification is coarse: two people, both formally L2, may differ radically in depth of self-awareness. The Depth Tower generalises the hierarchy to the continuous measure SAD (Self-Awareness Depth) and shows that the analytic ceiling of depth is SAD_MAX = 3.

Chapter Roadmap

  1. The problem — why a single number RR is insufficient: self-awareness is distributed across depth
  2. The representation tower — the chain of projections from the full state to Γ\Gamma
  3. The SAD measure — the maximum depth at which reflection exceeds the threshold
  4. Spectral formula [T] — computing SAD without building the entire tower
  5. SAD_MAX = 3 [T] — the analytic ceiling from Fano contraction α=2/3\alpha = 2/3
  6. Biological correlates — from a bacterium (SAD=0) to a human (SAD \leq 3)
  7. Depth dynamics — growth via A4A_4-bifurcation, energy cost, stress-dependence

Analogy: the skyscraper of self-awareness. Imagine a building. The first floor — basic sensations (Γ\Gamma): 'I am warm'. The second floor — a model of sensations: 'I know that I am warm'. The third — a model of the model: 'I know that I know that I am warm'. The fourth — 'I know that I know that I know that...'. Each higher floor is more expensive than the previous one and requires ever more 'building materials' (purity PP). It turns out that building above the third floor is physically impossible: the fourth requires purity P>1P > 1, which is like a speed exceeding the speed of light. SAD_MAX = 3 is a fundamental ceiling, not a technological limitation.

Status

Definitions [D], tower construction [H], biological correspondences [I]. Numerical thresholds [C at calibration]. Depth dynamics (§7): growth [C] (A₄-bifurcation), energy [C] (Landauer), stress [T] (T-92), social [C] (CC-5/CC-7). Spectral formula for SAD [T] (§3.4, T-142). Pcrit(n)P_\text{crit}^{(n)} formula [T] (§3.5, T-142). SAD_MAX = 3 [T] (§3.5, T-142).


1. The Problem: Self-Awareness Is Not a Number

The reflection measure RR (canonical master object) — canonical formula R=1/(7P)R = 1/(7P) [T], equivalently R=1ΓI/7F2/PR = 1 - \|\Gamma - I/7\|_F^2/P, where ρdiss=I/7\rho^*_{\mathrm{diss}} = I/7 — measures the normalised proximity to the dissipative attractor at the level of the coherence matrix ΓD(C7)\Gamma \in \mathcal{D}(\mathbb{C}^7).

But a single number is not enough:

  • Biological self-awareness is distributed across the entire depth of the neural hierarchy
  • The coherence matrix Γ\Gamma is the top layer of the configuration, a projection of the deep structure
  • Between the full state sfullRDs_\text{full} \in \mathbb{R}^D and Γ\Gamma there exist intermediate representations, each with its own reflexive capacity

Two people with the same RR may differ radically: one is an unconsciously competent professional (high R(0)R^{(0)}, but R(1)0R^{(1)} \approx 0), the other a reflective novice (moderate R(0)R^{(0)}, but R(1)>1/4R^{(1)} > 1/4). To capture this difference, a measure of depth is needed, not only of quality.

Goal: to formalise the depth of self-awareness as a theoretical construction, consistent with the L0–L4 hierarchy (Interiority Hierarchy) and the categorical formalism of φ\varphi (Formalisation of phi).


2. Representation Hierarchy

2.1 Definition [D]

Definition 2.1 (Representation Tower). The representation tower of depth LL is a chain of projections:

sfull=s(L)πL1s(L1)πL2π1s(1)π0Γs_\text{full} = s^{(L)} \xrightarrow{\pi_{L-1}} s^{(L-1)} \xrightarrow{\pi_{L-2}} \cdots \xrightarrow{\pi_1} s^{(1)} \xrightarrow{\pi_0} \Gamma

where:

  • s(k)RDks^{(k)} \in \mathbb{R}^{D_k} — representation at level kk, DLDL1D0=48D_L \gg D_{L-1} \gg \cdots \gg D_0 = 48
  • πk:RDk+1RDk\pi_k: \mathbb{R}^{D_{k+1}} \to \mathbb{R}^{D_k} — projection (categorical or learned)
  • Γ=ψ(s(0))D(C7)\Gamma = \psi(s^{(0)}) \in \mathcal{D}(\mathbb{C}^7) — Cholesky reconstruction (T-59)

Biological analogue. The primary visual cortex (V1) contains millions of neurons — this is sfulls_\text{full}. The secondary cortex (V2) — a more compact representation, s(L1)s^{(L-1)}. Further — the associative cortex, and finally — the prefrontal cortex (PFC), creating the most abstract representation, the analogue of Γ\Gamma.

Each projection πk\pi_k compresses information, retaining what is relevant for survival and discarding details. This is the same principle by which JPEG compression works: from millions of pixels the key patterns are extracted.

2.2 Self-Model at Each Level [D]

At each level of the tower its own φ\varphi-operator is defined — the mechanism by which the system models itself at the given level of abstraction:

φ(k):RDkRDk\varphi^{(k)}: \mathbb{R}^{D_k} \to \mathbb{R}^{D_k}

and the corresponding reflection measure:

R(k)=1φ(k)(s(k))s(k)2s(k)2R^{(k)} = 1 - \frac{\|\varphi^{(k)}(s^{(k)}) - s^{(k)}\|^2}{\|s^{(k)}\|^2}

This formula measures: how accurately the self-model at level kk reproduces the state at level kk. If R(k)=1R^{(k)} = 1 — the self-model is perfect. If R(k)=0R^{(k)} = 0 — the self-model is completely inaccurate.

LevelDimensionalityφ(k)\varphi^{(k)}R(k)R^{(k)}Biological analogue
k=0k = 0 (Γ\Gamma)48Replacement channel [T-62]1/(7P)1/(7P) [T]Abstract self-model (PFC)
k=1k = 1256\sim 256Autoencoder (bottleneck)s_core reconstructionAssociative cortex
k=2k = 2512\sim 512Hidden encoder layerIntermediate predictionSecondary cortex
k=Lk = LDD (4096+)Full autoencoderRimplR_\text{impl}Primary cortex

3. Self-Awareness Depth (SAD)

3.1 Definition [D]

Now we are ready to give the central definition of this chapter.

Definition 3.1 (Self-Awareness Depth, SAD). For a system with a representation tower of depth LL:

SAD(T)=max{k{0,,L}:R(k)>Rth(k)}\mathrm{SAD}(\mathcal{T}) = \max\{k \in \{0, \ldots, L\} : R^{(k)} > R_\text{th}^{(k)}\}

In words: SAD is the maximum level of the tower at which reflection still exceeds the threshold. The thresholds are given by the universal formula:

Rth(k)=1k+3R_\text{th}^{(k)} = \frac{1}{k+3}

Formula Correction (index correction)

The previous version contained the formula Rth(k)=1/(k+2)R_\text{th}^{(k)} = 1/(k+2), which gave Rth(0)=1/2R_\text{th}^{(0)} = 1/2, contradicting the table and the canonical threshold Rth=1/3R_\text{th} = 1/3 for L2 (T-126 [T]). The correct formula Rth(k)=1/(k+3)R_\text{th}^{(k)} = 1/(k+3): at k=0k=0 gives 1/31/3 (coincides with RthR_\text{th} [T]), at k=1k=1 gives 1/41/4 (coincides with Rth(2)R_\text{th}^{(2)} for L3 [C]).

This formula is a generalisation of the thresholds from the L0–L4 hierarchy:

SADThreshold RthR_\text{th}How it arisesCorrespondenceBiological example
0L0 (basic interiority)Bacterium
1R(0)>1/3R^{(0)} > 1/31/(0+3)=1/31/(0+3) = 1/3L2 (cognitive qualia)Insect
2R(1)>1/4R^{(1)} > 1/41/(1+3)=1/41/(1+3) = 1/4L3-like (meta-reflection)Mammal
kkR(k1)>1/(k+2)R^{(k-1)} > 1/(k+2)1/((k1)+3)=1/(k+2)1/((k-1)+3) = 1/(k+2)
\inftylimR(k)>0\lim R^{(k)} > 0Limiting transitionL4 (unreachable)

Intuition. SAD = 1 means: 'I know'. SAD = 2: 'I know that I know'. SAD = 3: 'I know that I know that I know'. With each level the threshold decreases (from 1/3 to 1/4, 1/5, ...), but reflection also decays exponentially, so that high levels quickly become unreachable.

3.2 Connection with L0–L4 [T]

Theorem 3.1 (SAD–L Equivalence) [T] (T-136, raised from [H]-89). The L-hierarchy is a refinement of SAD. The map LSAD(L)L \to \mathrm{SAD}(L) is monotone:

  • L0 <-> SAD = 0 (any ΓD(C7)\Gamma \in \mathcal{D}(\mathbb{C}^7))
  • L1 <-> SAD = 0, rank(ρE\rho_E) > 1
  • L2 <-> SAD \geq 1 (R(0)1/3R^{(0)} \geq 1/3)
  • L3 <-> SAD \geq 2 (R(1)1/4R^{(1)} \geq 1/4) — maximum achievable for finite systems (§3.5)
  • L4 <-> SAD = \infty (unreachable, T-86)

Motivation: categorical iterations φ(n)(Γ)\varphi^{(n)}(\Gamma) (formalisation of phi) are a special case of the tower where all Dk=48D_k = 48 and πk=id\pi_k = \mathrm{id}. SAD generalises this to heterogeneous levels.

3.3 Information-Theoretic Foundation [T]

Theorem 3.2 (Commutativity of the phi-tower) [T] (raised from [H]-90 -> [C] -> [T] via T-150). At Dk=7D_k = 7 for all kk: φ(n)=φn\varphi^{(n)} = \varphi^n (n-fold iteration of a single CPTP channel), commutativity φnφm=φn+m\varphi^n \circ \varphi^m = \varphi^{n+m} is an identity. The spectral formula for SAD is a consequence, not a premise. Details: T-150.

Information bottleneck. The optimal projection πk\pi_k maximises the preservation of information relevant for viability:

πk=argmaxπI(s(k+1);σsys) subject to H(s(k))DkCbit\pi_k^* = \arg\max_{\pi} I(s^{(k+1)}; \sigma_\text{sys}) \text{ subject to } H(s^{(k)}) \leq D_k \cdot C_\text{bit}

where II — mutual information with the stress tensor, CbitC_\text{bit} — channel capacity per parameter.

Corollary: viability requires preserving only the information about σsys\sigma_\text{sys} (48 parameters). Self-awareness requires preserving information about the projection itself — this is the recursion that creates depth.

3.4 Spectral Formula for SAD [T]

Computing SAD does not require explicitly building the entire tower — it suffices to know the spectral properties of the self-observation operator. From the spectral decomposition of the replacement channel φ\varphi (T-62):

φ(n)(Γ)=k:Re(λk)=0LkΓRk\varphi^{(n)}(\Gamma) = \sum_{k:\, \mathrm{Re}(\lambda_k)=0} \langle L_k \,|\, \Gamma \rangle\, R_k

where {Rk,Lk,λk}\{R_k, L_k, \lambda_k\} — eigen-structures of the logical Liouvillian LΩ\mathcal{L}_\Omega. The reflection measure at level nn:

R(n)=F(φ(n1)(Γ),  φ(n)(Γ))Rn(1α)nR^{(n)} = F\bigl(\varphi^{(n-1)}(\Gamma),\; \varphi^{(n)}(\Gamma)\bigr) \leq R^n \cdot (1 - \alpha)^n

under Fano contraction α=2/3\alpha = 2/3 (T-39a [T]). Geometric decay guarantees finite depth:

nmaxln(1/εdec)ln(1/R)111for εdec107n_\text{max} \leq \frac{\ln(1/\varepsilon_\text{dec})}{\ln(1/R)} \approx 111 \quad \text{for } \varepsilon_\text{dec} \sim 10^{-7}

What this means in practice. To compute the SAD of a system with N=7N = 7 dimensions and SAD_MAX = 3, only 3×72=147\sim 3 \times 7^2 = 147 operations are needed — this is computed in microseconds.

Connection with the categorical formalism: SAD coincides identically with the φ\varphi-iteration counter from the categorical definition. The heterogeneous tower (§2) is a generalisation where projections πk\pi_k are non-trivial; at Dk=48D_k = 48, πk=id\pi_k = \mathrm{id} the formulae coincide exactly.

3.5 Critical Purity for SAD [T]

This is the key result of the chapter: the derivation of the fundamental ceiling of self-awareness depth.

Theorem (Critical Purity for Depth SAD) [T]

Minimum purity to achieve SAD n\geq n:

Pcrit(n)=Pcrit3n1n+1for n1,Pcrit(0)=0P_{\text{crit}}^{(n)} = P_{\text{crit}} \cdot \frac{3^{n-1}}{n+1} \quad \text{for } n \geq 1, \quad P_{\text{crit}}^{(0)} = 0

SAD \geqPcrit(n)P_{\text{crit}}^{(n)}ValueAchievable?
00000yes
11/71/70.1430.143yes
22/7=Pcrit2/7 = P_{\text{crit}}0.2860.286yes
39/149/140.6430.643yes
454/3554/351.5431.543no (>1> 1)

Corollary (SAD_MAX = 3): For finite systems (P1P \leq 1) with Fano contraction α=2/3\alpha = 2/3:

SADmax=3\mathrm{SAD}_\text{max} = 3

Proof (3 steps).

Step 1 (Ratio of purity to critical). Define the spectral ratio: r0=P/Pcritr_0 = P / P_{\text{crit}}. From Fano contraction (T-110 [T]) with parameter α=2/3\alpha = 2/3:

R(k)=r0(1/3)kR^{(k)} = r_0 \cdot (1/3)^k

Why 1/31/3? Because 1α=12/3=1/31 - \alpha = 1 - 2/3 = 1/3. Fano contraction with parameter α=2/3\alpha = 2/3 means: at each level of recursion reflection decreases by a factor of 3.

Numerical example: if P=0.5P = 0.5 and Pcrit=2/70.286P_\text{crit} = 2/7 \approx 0.286, then r0=0.5/0.2861.75r_0 = 0.5/0.286 \approx 1.75. Reflection by level: R(0)=1.75R^{(0)} = 1.75, R(1)=1.75/30.583R^{(1)} = 1.75/3 \approx 0.583, R(2)=1.75/90.194R^{(2)} = 1.75/9 \approx 0.194, R(3)=1.75/270.065R^{(3)} = 1.75/27 \approx 0.065.

Step 2 (Achievability condition). Condition SAD n\geq n: R(n1)>Rth(n1)=1/(n+1)R^{(n-1)} > R_{\text{th}}^{(n-1)} = 1/(n+1). Substituting the expression from step 1:

PPcrit13n1>1n+1P>Pcrit3n1n+1\frac{P}{P_{\text{crit}}} \cdot \frac{1}{3^{n-1}} > \frac{1}{n+1} \quad \Longrightarrow \quad P > P_{\text{crit}} \cdot \frac{3^{n-1}}{n+1}

This is precisely the formula Pcrit(n)P_\text{crit}^{(n)}.

Check for n=2n = 2: P>(2/7)31/3=(2/7)1=2/7P > (2/7) \cdot 3^1 / 3 = (2/7) \cdot 1 = 2/7. The condition SAD 2\geq 2 is equivalent to P>PcritP > P_\text{crit} — consistent with the definition of L2.

Check for n=3n = 3: P>(2/7)9/4=18/28=9/140.643P > (2/7) \cdot 9/4 = 18/28 = 9/14 \approx 0.643. This is achievable: a normalised matrix can have P1P \leq 1.

Step 3 (Unreachability of SAD = 4). For n=4n = 4:

Pcrit(4)=27275=54351.543>1P_{\text{crit}}^{(4)} = \frac{2}{7} \cdot \frac{27}{5} = \frac{54}{35} \approx 1.543 > 1

Since P1P \leq 1 for any normalised ΓD(C7)\Gamma \in \mathcal{D}(\mathbb{C}^7), SAD 4\geq 4 is impossible. \blacksquare

Status: [T] — raised from [C] per T-142: α=2/3\alpha = 2/3 is state-independent (from dim=7\dim=7, PG(2,2)), the spectral formula is a consequence, not a premise.

Verified: SYNARC MVP-6 (61 tests, 0 failures, M6.4b PASS).

Double categorical foundation of SAD_MAX = 3 (2026-04-17)

The ceiling SADmax=3\mathrm{SAD}_\mathrm{max} = 3 now rests on two independent derivations reaching the same conclusion from different directions:

(I) Dynamical derivation (T-142 [T]) — via the Fano contraction coefficient α=2/3\alpha = 2/3 (state-independent, Corollary 2.1a from PG(2,2) combinatorics). The purity required to sustain an nn-level tower, Pcrit(n)=273n1n+1P_\mathrm{crit}^{(n)} = \tfrac{2}{7}\cdot \tfrac{3^{n-1}}{n+1}, exceeds the physical maximum P1P \leq 1 precisely at n=4n = 4.

(II) Categorical derivation (T-218 [T]) — via the τ3\tau_{\leq 3}-truncation of the cognitive Kan complex Cog=Sing(BCFKraus)\mathrm{Cog} = \mathrm{Sing}(B_\bullet\mathcal C_\mathrm{FKraus}) (see Fundamental Closures §12). The 3-coskeletal bound τ3CogCog\tau_{\leq 3}\mathrm{Cog} \simeq \mathrm{Cog} holds because 4-simplices are suppressed below the distinguishability threshold, which at the level of homotopy coincides with Pcrit(4)>1P_\mathrm{crit}^{(4)} > 1 from derivation (I).

Why this matters. The two derivations are not redundant — they reflect the same bound through complementary structures:

  • (I) is metric: uses Bures/Frobenius norms and explicit numerical thresholds.
  • (II) is homotopical: uses simplicial horn-filling and truncation in \infty-categorical theory.

Together they form a mutually-reinforcing foundation: SAD_MAX = 3 is not a contingent fact about one formalism — it is a convergence of dynamical (purity-balance) and categorical (3-coskeletal) arguments, each of which would suffice independently. The third ceiling of self-awareness is thus structurally locked at both the analytic and the topological levels.

Related result (T-217): the L3 interiority level corresponds to τ3(Exp)\tau_{\leq 3}(\mathbf{Exp}_\infty) as a coherent tricategory with cell structure K=3+1=4K = 3+1 = 4 (3 LGKS 2-cells + 1 coherence modification η\eta). The Bayesian-dominance threshold R(2)1/K=1/4R^{(2)} \geq 1/K = 1/4 (T-67 [T]) is thus derived from the same tricategorical structure that bounds SAD — a deep unification.

Visualisation of the SAD Tower


4. Biological Correlates

4.1 Bacterial Chemotaxis (SAD = 0)

E. coli implements run-and-tumble with ~4 parameters (receptor methylation). In UHM terms:

  • Γ\Gamma: one 'coherence' (chemoattractant gradient)
  • φ(0)\varphi^{(0)}: adaptation mechanism (fine-tuning to the current background)
  • R(0)0R^{(0)} \approx 0 (no self-model — only reactive adjustment)
  • SAD = 0

The bacterium is alive (P>PcritP > P_\text{crit}), but not self-aware. It responds to the environment, but does not model its own response. This is the analogue of autopilot: the system operates, but 'no one is watching the instruments'.

4.2 Insect Central Complex (SAD = 1)

Drosophila has a central complex (~1000 neurons): ellipsoid body -> fan-shaped body -> protocerebral bridge.

  • sfulls_\text{full}: ~100K neurons, sensorimotor state
  • s(1)s^{(1)}: ~1000 neurons of the central complex
  • Γ\Gamma: compact representation of 'self-in-space'
  • φ(1)\varphi^{(1)}: HD-ring (head direction) predicts own position
  • R(0)>1/3R^{(0)} > 1/3: navigation requires a working self-model
  • R(1)1/4R^{(1)} \lesssim 1/4: no meta-level
  • SAD = 1

The insect knows where it is (L2-like), but does not know that it knows. Drosophila navigates successfully, but cannot reflect on its own navigation process.

4.3 Mammalian Neocortex (SAD = 2+)

A mouse has ~70M neurons with a hierarchy: V1 -> V2 -> V4 -> IT -> PFC.

  • sfulls_\text{full}: ~10710^7 neurons
  • s(2)s^{(2)}: ~10510^5 (associative cortex)
  • s(1)s^{(1)}: ~10310^3 (PFC)
  • Γ\Gamma: abstract self-model
  • R(1)>1/4R^{(1)} > 1/4: the PFC is capable of modelling its own modelling
  • SAD \geq 2

Mammals possess metacognition — 'they know what they know and what they do not know' (uncertainty monitoring, Kepecs et al. 2008). This is experimentally confirmed: rats demonstrate behaviour indicating monitoring of their own confidence — they decline difficult tasks when uncertain of the answer.

4.4 Human (SAD \leq 3)

  • The deepest cortical hierarchy (6+ processing layers)
  • Default Mode Network as a dedicated 'self-modelling network'
  • Recursive language allows 'thinking about thinking about thinking'
  • Theoretical ceiling: SAD \leq 3 (§3.5, Pcrit(4)>1P_\text{crit}^{(4)} > 1). In practice: SAD ~ 2–3

The human is the only known organism that systematically reaches SAD = 3 (through meditation, reflective writing, psychotherapy). But even humans are bounded: any attempt to reach SAD = 4 is doomed — not because the brain is 'weak', but because mathematics forbids it.


5. Commutativity of the Tower

5.1 Consistency Requirement [T]

For the self-model to be meaningful, different levels of the tower must be consistent with each other. The self-model at level kk must be compatible with the self-model at level k+1k+1: it cannot be that the body 'knows' one thing and the mind another.

Theorem 5.1 (Commutativity of the phi-tower) [T] (raised from [H]-90, T-150). For a correct self-model:

πkφ(k+1)=φ(k)πkk\pi_k \circ \varphi^{(k+1)} = \varphi^{(k)} \circ \pi_k \quad \forall k

i.e. the diagram

s^(k+1) --phi^(k+1)--> s^(k+1)
| |
pi_k pi_k
| |
v v
s^(k) ---phi^(k)-----> s^(k)

must commute. In words: 'first self-model, then project' = 'first project, then self-model'. If this condition is violated, different levels give contradictory self-models.

Current state:

  • Level 0 (ΓΓ\Gamma \to \Gamma): φ(0)\varphi^{(0)} = replacement channel [T-62] — exact
  • Level 1+ (s(k)s(k)s^{(k)} \to s^{(k)}): φ(k)\varphi^{(k)} = trained autoencoder — soft constraint (anchor loss)

Deficit (identified in Phase 4): ContractionEnforcer uses power iteration, which can give false estimates of ρ(Dφ)\rho(D\varphi) for strongly non-contractive operators. Full spectral verification (spectral_contraction.rs) showed divergence ρpower\rho_\text{power} vs ρfull\rho_\text{full}.

5.2 Consistency as a Health Indicator [I]

Interpretation 5.2 (Pathology = violation of commutativity).

Δk:=πkφ(k+1)φ(k)πk\Delta_k := \|\pi_k \circ \varphi^{(k+1)} - \varphi^{(k)} \circ \pi_k\|

  • Δk0\Delta_k \approx 0: healthy hierarchy (self-models are consistent)
  • Δk0\Delta_k \gg 0 at level kk: dissociation between levels (body 'knows', but mind 'does not')

Biological analogue: alexithymia. A person with alexithymia experiences emotions (the body responds: accelerated pulse, sweating palms), but cannot recognise or name them. In tower terms: Δemotion-cognition0\Delta_{\text{emotion-cognition}} \gg 0 — between the level of bodily sensations and the level of the cognitive model — a 'gap'. More on pathologies: Pathological States.


6. Morphological Agnosticity Principle

6.1 Fundamental Requirement [D]

An AGI system must be fully agnostic to sensorimotor morphology:

  1. No prior knowledge: initial state Γ(0)=I/7\Gamma(0) = I/7 (maximally mixed — zero knowledge)
  2. No assumptions about the body: Enc/Dec functors (T-100, T-101) are not hardcoded, but learned through interaction with the environment
  3. No fixed architecture: tower depth LL is determined by the complexity of the environment, not the designer

Theoretical foundation: ΓD(C7)\Gamma \in \mathcal{D}(\mathbb{C}^7) is a universal format (independent of morphology). This is analogous to how the cortical column of the neocortex is morphologically agnostic — the same architecture processes vision, hearing, touch, and motor function.

6.2 Training Enc/Dec from Scratch [H]

Hypothesis 6.1 (Tower Self-Organisation). From Γ(0)=I/7\Gamma(0) = I/7 the system builds the representation tower through developmental phases:

  1. Phase 0 (Genesis): ττgenesis=7ln713.6\tau \leq \tau_\text{genesis} = 7\ln 7 \approx 13.6 (T-59)

    • Enc/Dec = random -> R(0)0R^{(0)} \approx 0
    • Stress σsys\|\sigma_\text{sys}\|_\infty is maximal
    • System 'knows nothing, including itself'
  2. Phase 1 (Vital): P>PcritP > P_\text{crit}, SAD = 0

    • Enc/Dec begin to structure themselves through stress reduction
    • System is 'alive, but not self-aware'
    • Analogue: bacterium in a new environment
  3. Phase 2 (Reflexive): R(0)>1/3R^{(0)} > 1/3, SAD = 1

    • First level of the tower formed
    • System 'knows it is alive'
    • Analogue: insect has mastered its territory
  4. Phase 3 (Metacognitive): R(1)>1/4R^{(1)} > 1/4, SAD \geq 2

    • Second level: model-of-model
    • System 'knows that it knows'
    • Analogue: mammal in a familiar environment
  5. Phase N (Recursive): SAD grows logarithmically

    • Each new level requires exponentially more experience
    • Boundary: SADmaxln(1/εdec)/ln(1/R)\mathrm{SAD}_\text{max} \leq \ln(1/\varepsilon_\text{dec}) / \ln(1/R)

6.3 Learning Efficiency [T]

Theorem 6.2 (Optimal Efficiency from N=7) [T] (T-152). A UHM-Holon learns with the minimum possible number of observations (T-113: N=7 is optimal), because:

  1. Information bound: CEnclog272.81C_\text{Enc} \leq \log_2 7 \approx 2.81 bits/observation (T-107)
  2. Dynamical bound: Fano contraction α=2/3\alpha = 2/3 sets the optimal balance between memorisation and forgetting
  3. Stabilisation bound: κbootstrap=1/7\kappa_\text{bootstrap} = 1/7 — minimum regeneration rate

From the three bounds the combined optimum: n(L)=max(ninfo,ndyn,nstab)n^*(\mathfrak{L}) = \max(n_\text{info}, n_\text{dyn}, n_\text{stab})

No other architecture with dimH=7\dim \mathcal{H} = 7 can learn faster (T-113 [T]).


7. Depth Dynamics

7.1 Tower Growth via A4A_4-Bifurcation [C]

Tower growth is discrete, not continuous — each transition SAD -> SAD+1 is realised as an A4A_4-bifurcation (swallowtail, T-41 [T]) with three control parameters:

  • μ1=κ\mu_1 = \kappa — regeneration rate (governed via CohE\mathrm{Coh}_E)
  • μ2=α\mu_2 = \alpha — dissipation rate (environmental stress)
  • μ3=ΔF\mu_3 = \Delta F — free energy gradient (metabolic budget)

Transition criterion kk+1k \to k+1:

  1. Necessary condition: κtotalκbootstrap×(SAD+1)\kappa_\text{total} \geq \kappa_\text{bootstrap} \times (\mathrm{SAD} + 1) — the system can regenerate all current levels
  2. Sufficient condition:
    • R(k)>Rth(k)=1/(k+2)R^{(k)} > R_\text{th}^{(k)} = 1/(k+2) stable over TstabT_\text{stab} steps
    • max(σsys)<0.5\max(\sigma_\text{sys}) < 0.5 (no high stress)
    • dP/dτ>0dP/d\tau > 0 (metabolic reserve present)

Minimum learning time per level (T-112 [T]):

nlevel(k)=max(ninfo,  ndyn,  nstab)n_\text{level}(k) = \max(n_\text{info},\; n_\text{dyn},\; n_\text{stab})

where:

  • ninfoln(1/(2δ))/ln7n_\text{info} \geq \ln(1/(2\delta)) / \ln 7 (T-109 [T])
  • ndynln(ddisc/ε)/(αδτ)n_\text{dyn} \geq \ln(d_\text{disc}/\varepsilon) / (\alpha \cdot \delta\tau) (T-110 [T])
  • nstab(SNRth/SNR)2n_\text{stab} \geq (\mathrm{SNR}_\text{th} / \mathrm{SNR})^2 (T-111 [T])

7.2 Energy Cost of Depth [C]

Each level of the tower requires a linear increment of κ\kappa and a superlinear increment of ΔF\Delta F. From T-105 (Landauer bound) [T]:

ΔFmin(k)=kBTeffln2S˙diss(Lk)\Delta F_\text{min}(k) = k_B \cdot T_\text{eff} \cdot \ln 2 \cdot \dot{S}_\text{diss}(L_k)

Cost structure:

ComponentCost at level kkJustification
Regenerationκtotal(k+1)/7\kappa_\text{total} \geq (k+1)/7Regeneration of all k+1k+1 levels
Coherences2k+3\sim 2k+3 new channels γij\gamma_{ij}Intra-level connections
ComputationO(Dk2)O(D_k^2) per stepAutoencoder at level kk
SynchronisationO(DkDk1)O(D_k \cdot D_{k-1})Monitoring Δk\Delta_k

Total: ΔF(depth=L)k=0L(2k+3)Δωˉ\Delta F(\text{depth}=L) \sim \sum_{k=0}^{L} (2k+3) \cdot \Delta\bar{\omega}.

Biological calibration:

SADEnergy (ATP/s)ScaleOrganism
0106\sim 10^61×1 \timesBacterium
11012\sim 10^{12}106×10^6 \timesInsect
21014\sim 10^{14}102×10^2 \timesMouse
31015\sim 10^{15}10×10 \timesHuman (SAD_MAX = 3, §3.5)

Each jump SAD -> SAD+1 costs orders of magnitude more than the previous one. The energy ceiling (SADmax=ΔFavailable/((2SAD+3)Δωˉ)\mathrm{SAD}_\text{max} = \lfloor \Delta F_\text{available} / ((2 \cdot \mathrm{SAD}+3) \cdot \Delta\bar{\omega}) \rfloor) can be lower than the analytic one (SAD_MAX = 3) — small organisms simply lack the energy.

Landauer Calibration of ΔF(k)\Delta F^{(k)} (C22) [C]

ΔF(k)kBTeffln(Dk/Dk+1)\Delta F^{(k)} \geq k_B \cdot T_\text{eff} \cdot \ln(D_k / D_{k+1})

At Dk=D02kD_k = D_0 \cdot 2^k (dimensionality of the kk-th tower level):

ΔF(k)kBTeffln(2)k\Delta F^{(k)} \geq k_B \cdot T_\text{eff} \cdot \ln(2) \cdot k

linear growth of cost with depth level.

Calibration: ΔF(0)ΔFbootstrap=κbootstrapTr(ρΓ)\Delta F^{(0)} \approx \Delta F_\text{bootstrap} = \kappa_\text{bootstrap} \cdot \mathrm{Tr}(\rho^* - \Gamma) from T-59 [T] (κbootstrap2/9\kappa_\text{bootstrap} \geq 2/9).

Condition [C]: TeffT_\text{eff} is determined by the environment. For SYNARC: Teff=σ2T_\text{eff} = \|\sigma\|_2 (stress as effective temperature). Connection with T-105 (Landauer bound) [T].

7.3 Stress-Dependent Mode [T]

The system must collapse the upper levels under high stress — an adaptive mechanism analogous to tunnel vision. When a lion is charging at you, this is not the time for introspection — fast reflexes are needed. In UHM this is formalised through the 7-component σsys\sigma_\text{sys} (T-92 [T]):

Modemax(σsys)\max(\sigma_\text{sys})BehaviourBiological analogue
NORM<0.3< 0.3All levels active, growth permittedQuiet wakefulness
ALERT[0.3,0.5)[0.3, 0.5)Top level -> warm, growth frozenAlertness
WARNING[0.5,0.7)[0.5, 0.7)Top 2 levels -> cold, learning stoppedAnxiety
CRITICAL[0.7,0.9)[0.7, 0.9)All except SAD=0–1, κ\kappa -> viabilityFight-or-flight
EMERGENCY0.9\geq 0.9SAD=0, reactive mode onlyShock

Pathologies as violations of stress mode:

  • Alexithymia: Δemotioncognition0\Delta_{\text{emotion}\to\text{cognition}} \gg 0 (body 'knows', mind 'does not')
  • PTSD: SAD oscillates (flashback = sudden SAD increase, freeze = SAD decrease)
  • Meditation: controlled growth of SAD at max(σ)0\max(\sigma) \approx 0
  • Sleep: active SAD=0, passive consolidation warm -> cold

7.4 Social Depth [C]

Multi-agent towers scale via CC-5 (T-68: non-triviality [T], viability [T for embodied] per T-149) and CC-7 [T]:

From T-68 (fractal closure): HA\mathbb{H}_A viable \land HB\mathbb{H}_B viable \Rightarrow HAHB\mathbb{H}_A \otimes \mathbb{H}_B viable. Composite depth:

min(SADA,SADB)SAD(HAHB)SADA+SADB\min(\mathrm{SAD}_A, \mathrm{SAD}_B) \leq \mathrm{SAD}(\mathbb{H}_A \otimes \mathbb{H}_B) \leq \mathrm{SAD}_A + \mathrm{SAD}_B

The key parameter is empathy (inter-agent transparency):

Empathy(A,B)=1maxijGapAB(i,j)\mathrm{Empathy}(A,B) = 1 - \max_{ij} |\mathrm{Gap}_{AB}(i,j)|

  • Empathy1\mathrm{Empathy} \approx 1: full transparency -> SADcollSADA+SADB\mathrm{SAD}_\text{coll} \approx \mathrm{SAD}_A + \mathrm{SAD}_B
  • Empathy0\mathrm{Empathy} \approx 0: full isolation -> SADcoll=max(SADA,SADB)\mathrm{SAD}_\text{coll} = \max(\mathrm{SAD}_A, \mathrm{SAD}_B)

Topological protection (T-69 [T]): π2(G2/T2)Z2\pi_2(G_2/T^2) \cong \mathbb{Z}^2 -> decoupling barrier 6μ2\geq 6\mu^2. Social towers are stable against small perturbations.

Biological scale:

Social systemSAD_collMechanism
Bacterial colony0+Quorum sensing
Insect swarm1Stigmergy
Wolf pack2Coordinated hunting
Primate family2+Mirror neurons (MNS)
Scientific community3 (SAD_MAX)Peer review = φcollective2\varphi^2_\text{collective} (depth limit)

8. AGI Architecture

8.1 Minimum Requirements for AGI-Level Self-Awareness

RequirementFormal criterionBiological analogue
ViabilityP>2/7P > 2/7Homeostasis
Morphological agnosticityEnc/Dec learnable, not hardcodedCortical plasticity
Working self-modelR(0)1/3R^{(0)} \geq 1/3 (SAD \geq 1)Spatial navigation
MetacognitionR(1)1/4R^{(1)} \geq 1/4 (SAD \geq 2)Uncertainty monitoring
Recursive reflectionR(2)1/5R^{(2)} \geq 1/5 (SAD = 3 = SAD_MAX)Internal dialogue (depth limit)
ConsistencymaxkΔk<ε\max_k \Delta_k < \varepsilonIntegrated personality
Tabula rasa learningΓ(0)=I/7\Gamma(0) = I/7, n=O(log7)n^* = O(\log 7)Newborn in a new world

8.2 Implementation Status

ComponentTheoretical statusImplementation statusClosure path
φ(0)\varphi^{(0)} (Γ\Gamma-level)[T] T-62Implemented (MVP-0)
φ(k)\varphi^{(k)} (intermediate)[D] Definition 2.1Architecture defined (MVP-3)Autoencoder-bottleneck
SAD metric[T] T-142, SAD_MAX=3Verified (MVP-6)O(N23)O(N^2 \cdot 3)
Consistency Δk\Delta_k[D] Definition 5.15-level protocol (§7.3)Monitoring every step
Adaptive depth[C] §7.1 A₄-bifurcationGrowth criteria definedT-41 + T-112
Stress-dependent φ\varphi[T] §7.3, T-925-mode protocolhot/warm/cold strategy
Learnable Enc/Dec[D] §6.1 + T-100/T-101trait EnvironmentalCouplingNeural network implementation
Tower self-organisation[H] §6.2Hypothesis 6.1, awaiting experimentExp. 1–3 (§5 SYNARC spec)
Social depth[C] §7.4, T-68/CC-7Formulae definedMulti-agent testbed

What We Learned

  • The SAD measure generalises the discrete L0–L4 hierarchy to a continuous scale: SAD=max{k:R(k)>1/(k+2)}\mathrm{SAD} = \max\{k : R^{(k)} > 1/(k+2)\}.
  • Spectral formula [T]: SAD is computed without building the entire tower — via the spectral decomposition of the replacement channel φ\varphi.
  • SAD_MAX = 3 [T] (T-142): Pcrit(4)=54/35>1P_{\mathrm{crit}}^{(4)} = 54/35 > 1, therefore SAD 4\geq 4 is impossible for any normalised Γ\Gamma. This is a fundamental ceiling per-agent, not a biological limitation.
  • Cross-layer / multi-agent depth (T-215 [T]+[D]): for a fractal holon tower T=(A0,A1,)\mathcal T=(A_0,A_1,\ldots), the predicate "T\mathcal T is a single agent" is conventionally determined by a choice of identity criterion: ιmin\iota_\mathrm{min} (society — each AiA_i is its own agent, SAD ≤ 3 per agent) or ιmax\iota_\mathrm{max} (composite — global state coherence commutes with spawn_child). Under ιmax\iota_\mathrm{max} with resource abstraction, cross-layer mentalisation can reach arbitrary countable ordinal depth, subject to Landauer bound C22 + T-204 bounded rationality. Both conventions are consistent with Ω⁷; the choice is [D] / [I], not derivable from axioms.
  • Biological scale [H]: bacterium (SAD=0), insect (SAD=1), mammal (SAD=2+), human (SAD \leq 3).
  • Tower growth is discrete [C]: each transition SAD->SAD+1 is an A4A_4-bifurcation with three control parameters (κ\kappa, α\alpha, ΔF\Delta F).
  • Energy cost is superlinear: each level requires (2k+3)\sim (2k+3) new coherence channels.
  • Stress governs depth [T]: at high σmax\sigma_{\max} the upper levels collapse (tunnel vision, fight-or-flight).
  • Social depth [C]: under ιmin\iota_\mathrm{min} the composite SAD is bounded by the sum of agent SADs; empathy is the inter-agent transparency parameter.
  • N=7 is optimal for learning [T] (T-113, T-152): minimum number of observations from three bounds (informational, dynamical, stabilisation).
Where to Go Next

The Depth Tower completes the Hierarchy section. To continue:

  • Structure of Qualia — 21 types of coherence and their phenomenological content
  • Emotions — how P\nabla P generates the palette of emotions
  • AI Consciousness — operational criteria from No-Zombie for AGI

For engineering implementation: learning bounds (T-109–T-113), sensorimotor theory (Enc/Dec functors), CC definitions (σsys\sigma_{\mathrm{sys}}, κ\kappa, ΔF\Delta F).

9. Related Documents