Skip to main content

Sensorimotor Theory

"A living being is not a thing but a process: a continuous exchange with the environment in which the boundary between 'self' and 'not-self' is recreated at every moment." — Francisco Varela

Who this chapter is for

The sensorimotor cycle as a consequence of the canonical evolution equation. The reader will learn how perception, evaluation, and action are derived from the dynamics of Γ\Gamma.

Every living being — from a bacterium feeling a chemical gradient to a human navigating an unfamiliar city — solves the same problem: perceive the environment and respond appropriately. This task seems mundane, but behind it lies one of the deepest problems in the science of complex systems.

Imagine an amoeba. It has no eyes, ears, or brain. Yet it distinguishes: it moves toward nutrients, avoids poison, flows around obstacles. Between "chemical signal on the membrane" and "extending a pseudopod" stands something — not a mere reflex, but a closed cycle: perception → evaluation → action → perception. This cycle — the sensorimotor loop — is the minimal unit of adaptive behaviour.

Classical control theory describes the sensorimotor loop as "sensor → controller → actuator". Active inference (FEP) sees it as minimising variational free energy. Reinforcement learning models it as maximising cumulative reward. Each of these approaches captures part of the truth — but none answers two key questions:

  1. Why is the cycle structured this way? Where do the number of perception channels, the action structure, and the format of internal evaluation come from?
  2. Where in the cycle does experience fit? When is the sensorimotor cycle accompanied by subjective experience, and when is it not?

Coherence Cybernetics (CC) gives a constructive answer to both questions. The sensorimotor cycle is not postulated — it is derived from the canonical 3-term evolution equation. The environment does not introduce a "fourth force": it modifies the three already existing channels (Hamiltonian, dissipative, regenerative). Experience turns out to be not a side effect but an integral part of the cycle — through the hedonic valence Vhed\mathcal{V}_{\text{hed}}, which guides action.

In this chapter we build a complete formal theory of sensorimotor encoding: from axiomatic grounding to concrete architectures and predictions. The reader familiar with the introduction and the evolution equation will find here a natural continuation — a step from "how the system lives within itself" to "how the system interacts with the world".

In the previous chapter we traced 80 years of cybernetics history and saw that each tradition — from Wiener to Friston — captured part of the sensorimotor problem: feedback, the observer, active inference. Now we show how CC solves it completely — without additional postulates, within the same 3-term evolution equation.

Chapter roadmap

In this chapter we:

  1. Prove that the environment does not add a 4th term — Theorem T-102 on completeness of the 3-channel decomposition (Section 1).
  2. Construct the perception functor Enc — how the environment enters the system through modification of the evolution equation (Section 2).
  3. Construct the action functor Dec — how the system chooses the optimal action through a min-max strategy (Section 3).
  4. Derive the hedonic mechanism — why pleasure and suffering are not side effects but derivatives of viability (Section 5).
  5. Classify 21 qualia-types as sensorimotor channels (Section 6).
  6. Establish fundamental limits — information capacity log27\leq \log_2 7 bits (T-107) and compositionality of Enc/Dec (T-108) (Sections 9–10).
  7. Compare with classical approaches — control theory, FEP, RL — as projections of CC (Section 14).
On notation

In this document:

  • Γ\Gammacoherence matrix
  • θij=arg(γij)\theta_{ij} = \arg(\gamma_{ij}) — coherence phases
  • σsys\sigma_{\mathrm{sys}}stress tensor (T-92 [T])
  • hext=h(H)+h(D)+h(R)h^{\text{ext}} = h^{(H)} + h^{(D)} + h^{(R)}3-channel decomposition [T]
  • P=Tr(Γ2)P = \mathrm{Tr}(\Gamma^2)purity
  • φ\varphiself-modelling operator
  • ρ=φ(Γ)\rho_* = \varphi(\Gamma)target state

This document describes the formal theory of sensorimotor encoding — how a holon perceives its environment and acts upon it, remaining within the canonical 3-term evolution equation.

Key result: the external force F_ext is not a 4th term of the evolution equation, but a modification of the three existing channels (Hamiltonian, dissipative, regenerative). The completeness of this decomposition is proven by the LGKS Theorem (T-57 [T]).


1. Canonical Inclusion of the Environment

1.1 The 3-term equation as closed dynamics

The evolution equation of a holon:

dΓdτ=i[Heff,Γ]+DΩ[Γ]+R[Γ,E]\frac{d\Gamma}{d\tau} = -i[H_{\text{eff}}, \Gamma] + \mathcal{D}_\Omega[\Gamma] + \mathcal{R}[\Gamma, E]

contains exactly three terms [T]:

TermTypeCanonical origin
i[Heff,Γ]-i[H_{\text{eff}}, \Gamma]Unitary (Hamiltonian)Axiom A3
DΩ[Γ]\mathcal{D}_\Omega[\Gamma]Dissipative (Lindblad)Liouvillian
R[Γ,E]\mathcal{R}[\Gamma, E]RegenerativeCategorical conjugation

1.2 The environment modifies 3 channels, not adds a 4th

Intuitively: imagine a violinist in an orchestra. The conductor influences them (suggesting tempo), the hall acoustics (blurring sound), and the other musicians (helping to return to the common key). These three types of influence are everything there is. There is no fourth type of influence on the violinist that would not be a combination of the conductor's gesture, acoustic noise, and adjustment to the ensemble. Theorem T-102 formalises exactly this intuition: the environment cannot "reach" a holon by any means other than the three canonical channels.

Theorem T-102 (Completeness of the 3-term equation) [T]

Statement

Any CPTP-compatible external influence on a holon decomposes into the sum of three channels:

hext=h(H)+h(D)+h(R)h^{\text{ext}} = h^{(H)} + h^{(D)} + h^{(R)}

where h(H)h^{(H)} modifies HeffH_{\text{eff}}, h(D)h^{(D)} modifies DΩ\mathcal{D}_\Omega, h(R)h^{(R)} modifies R\mathcal{R}. A fourth type of CPTP generator does not exist.

Proof. Direct consequence of the LGKS Theorem (T-57 [T], completeness of the triadic decomposition):

  1. An arbitrary generator of a CPTP semigroup L\mathcal{L} on D(C7)\mathcal{D}(\mathbb{C}^7) has the Gorini–Kossakowski–Sudarshan–Lindblad form:
L[ρ]=i[H,ρ]+k(LkρLk12{LkLk,ρ})\mathcal{L}[\rho] = -i[H, \rho] + \sum_k \left(L_k \rho L_k^\dagger - \tfrac{1}{2}\{L_k^\dagger L_k, \rho\}\right)
  1. Any external influence preserving CPTP properties of the dynamics is a perturbation LL+δL\mathcal{L} \to \mathcal{L} + \delta\mathcal{L}
  2. The perturbation δL\delta\mathcal{L} has the same LGKS form → decomposes into δH\delta H (Hamiltonian part) and δLk\delta L_k (Lindblad part)
  3. Triadic decomposition {Lk}\{L_k\}: dissipative + regenerative operators. A fourth type is forbidden by T-57. \blacksquare

Corollary: The F_ext term in simulation is not a separate force, but a composition of three modifications:

ChannelPerturbation formulaPhysical meaningExample
h(H)h^{(H)}δ(Δωij)\delta(\Delta\omega_{ij})Energetic coupling with environmentSensory input, neuromodulators
h(D)h^{(D)}δΓ2θ˙ij\delta\Gamma_2 \cdot \dot{\theta}_{ij}Environmental noiseStress, interference, temperature
h(R)h^{(R)}δκ(θijtargetθij)\delta\kappa \cdot (\theta^{\text{target}}_{ij} - \theta_{ij})Modification of regenerationMeditation, psychotherapy, learning

Canonical form — from Definition 8.1 [T]:

Lext=i<jhijextγijsin(θij)\mathcal{L}_{\text{ext}} = \sum_{i<j} h^{\text{ext}}_{ij} \cdot |\gamma_{ij}| \cdot \sin(\theta_{ij})

2. Perception Functor Enc

2.0 Intuition: what does it mean to "perceive"

What does the eye do when it sees an apple? From a physics perspective — it converts electromagnetic waves into neural impulses. From a computer-science perspective — it encodes input data into a latent representation. But both perspectives miss the essential point: perception is not passive recording, but active inclusion of the environment in the system's own dynamics.

When an amoeba "senses" glucose, its internal state changes — not because information was "recorded" somewhere, but because glucose molecules literally altered the dynamics of intracellular processes. Perception is a deformation of one's own equation of motion under the influence of the environment.

The Enc functor formalises exactly this: it maps an observation oo not into a "record in memory" but into a modification of the evolution equation — a triple (h(H),h(D),h(R))(h^{(H)}, h^{(D)}, h^{(R)}) that changes the Hamiltonian, dissipator, and regenerator. Perceiving an apple simultaneously changes the energy landscape (shape, colour), modifies the noise characteristics (texture, motion), and shifts the target state (hunger → satiation).

2.1 Definition

Theorem T-100 (Environmental encoding) [T]

Statement

For a holon H\mathbb{H} with coherence matrix ΓD(C7)\Gamma \in \mathcal{D}(\mathbb{C}^7), there exists a unique (up to G2G_2-gauge) CPTP encoding functor:

Enc:ObsSpaceEnd(D(C7))\mathrm{Enc}: \mathrm{ObsSpace} \to \mathrm{End}(\mathcal{D}(\mathbb{C}^7))

satisfying:

  1. CPTP: Enc(o)[Γ]\mathrm{Enc}(o)[\Gamma] is a state for any observation oObsSpaceo \in \mathrm{ObsSpace}
  2. 3-channel decomposition: Enc(o)=δH(o)δD(o)δR(o)\mathrm{Enc}(o) = \delta H^{(o)} \oplus \delta D^{(o)} \oplus \delta R^{(o)}
  3. Functoriality: Enc(o1o2)=Enc(o1)Enc(o2)\mathrm{Enc}(o_1 \circ o_2) = \mathrm{Enc}(o_1) \circ \mathrm{Enc}(o_2)

Proof.

  1. Existence: the environment acts through hijexth^{\text{ext}}_{ij} (Def. 8.1 [T]). The map ohext(o)o \mapsto h^{\text{ext}}(o) defines Enc\mathrm{Enc}.
  2. 3-channel: follows from T-102 (completeness of the 3-term equation).
  3. Uniqueness: consequence of G2G_2-rigidity (uniqueness theorem [T]) — for a system satisfying (AP)+(PH)+(QG)+(V), the map is unique up to G2=Aut(O)G_2 = \mathrm{Aut}(\mathbb{O}).
  4. Functoriality: CPTP channels are closed under composition. \blacksquare

Example: neuroscientific implementation. The visual cortex implements Enc hierarchically: V1 extracts edges (hAS(H)h^{(H)}_{AS} — articulation of structure), V4 encodes colour (hAE(H)h^{(H)}_{AE} — articulation of interiority), and MT encodes motion (hAD(D)h^{(D)}_{AD} — dissipative component of dynamics). All three channels converge in association areas, forming a unified hexth^{\text{ext}}. Functoriality guarantees that the scene "red ball moving left" is encoded identically whether perceived as a whole or in parts.

2.2 Implementation via 7 observable indices

The Γ measurement protocol defines 7 observable indices IiI_i (i{A,S,D,L,E,O,U}i \in \{A, S, D, L, E, O, U\}), each mapping to a specific component of hexth^{\text{ext}}:

IndexFormulaChannel hexth^{\text{ext}}Measurement
IAI_A (articulation)I(input;latent)/H(input)I(\text{input}; \text{latent}) / H(\text{input})hA,(H)h^{(H)}_{A,\cdot}Hamiltonian
ISI_S (structure)rankε(Jf)/min(dout,din)\mathrm{rank}_\varepsilon(J_f) / \min(d_{\text{out}}, d_{\text{in}})hS,(H)h^{(H)}_{S,\cdot}Hamiltonian
IDI_D (dynamics)maxiλiLyap\max_i \lambda_i^{\text{Lyap}} (normalised)hD,(D)h^{(D)}_{D,\cdot}Dissipative
ILI_L (logic)1[fi,fj]F/(fifj)1 - \|[f_i, f_j]\|_F / (\|f_i\| \cdot \|f_j\|)hL,(H)h^{(H)}_{L,\cdot}Hamiltonian
IEI_E (interiority)exp(SvN(ρattn))\exp(S_{vN}(\rho_{\text{attn}}))hE,(R)h^{(R)}_{E,\cdot}Regenerative
IOI_O (ground)1ϵhF1 - \|\nabla_\epsilon \mathbf{h}\|_FhO,(D)h^{(D)}_{O,\cdot}Dissipative
IUI_U (unity)Φeff=λ2(L)/λmax(L)\Phi_{\text{eff}} = \lambda_2(L) / \lambda_{\max}(L)hU,(R)h^{(R)}_{U,\cdot}Regenerative

Logic of channel assignment:

  • Hamiltonian h(H)h^{(H)}: informational indices (IA,IS,ILI_A, I_S, I_L) — alter the energy landscape, i.e., which states are more or less probable
  • Dissipative h(D)h^{(D)}: load indices (ID,IOI_D, I_O) — amplify/attenuate decoherence
  • Regenerative h(R)h^{(R)}: integrative indices (IE,IUI_E, I_U) — modulate the rate of recovery

Intuitively: these are three types of "sense organs". Hamiltonian indices are analytical senses (vision, hearing): they report what is happening in the environment, altering the internal preference landscape. Dissipative indices are load senses (fatigue, heat): they report how chaotic the environment is, amplifying internal noise. Regenerative indices are recovery senses (calm, safety): they report whether the environment supports self-healing.

2.3 Quasi-functor G

For AI systems, encoding is implemented through the quasi-functor G:AIStateD(C7)G: \mathrm{AIState} \to \mathcal{D}(\mathbb{C}^7), defined in the measurement protocol:

G(x)=argminΓD(C7)[Lreconstruct(Γ,{Ii(x)})+λphysLphys(Γ)]G(\mathbf{x}) = \arg\min_{\Gamma \in \mathcal{D}(\mathbb{C}^7)} \left[\mathcal{L}_{\text{reconstruct}}(\Gamma, \{I_i(\mathbf{x})\}) + \lambda_{\text{phys}} \cdot \mathcal{L}_{\text{phys}}(\Gamma)\right]

where Lphys\mathcal{L}_{\text{phys}} includes purity, spectral gap, and Cholesky decomposition constraints.


3. Action Functor Dec

3.0 Intuition: what does it mean to "act"

If Enc is "how the environment enters the system", then Dec is "how the system exits into the environment". But "acting" in CC is not just "sending a motor command". Action is the choice of that modification of the environment which minimises the largest deficit of internal resources.

Imagine a person who simultaneously has a headache and a rumbling stomach. Which action will they choose? If the headache is stronger — take a tablet. If the hunger is stronger — go eat. They do not minimise "average pain" (this would allow ignoring catastrophic channels), but eliminate the maximum deficit. This is exactly what the operator argminamaxkσkmotor\arg\min_a \max_k \sigma^{\mathrm{motor}}_k does — it guarantees that no channel ends up in a critical state.

Analogy with robotics: this is not a PID controller minimising error along one axis, nor a quadratic regulator minimising a weighted sum of errors. This is a min-max strategy — as in game theory, where the player chooses a move minimising the worst outcome.

3.1 Definition

Theorem T-101 (Optimal action) [T]

Statement

For a holon with current state Γ\Gamma and stress tensor σsys(Γ)\sigma_{\mathrm{sys}}(\Gamma) [T] (T-92), the optimal action is defined as:

a=argminaAσsys(Γ(τ+δτa))a^* = \arg\min_{a \in \mathcal{A}} \|\sigma_{\mathrm{sys}}(\Gamma(\tau + \delta\tau \mid a))\|_\infty

where Γ(τ+δτa)\Gamma(\tau + \delta\tau \mid a) is the predicted state under action aa, and \|\cdot\|_\infty is the sup-norm of the stress tensor.

Proof.

  1. Equivalence of viability conditions (T-92 [T]):
P(Γ)>27    σsys(Γ)<1P(\Gamma) > \frac{2}{7} \iff \|\sigma_{\mathrm{sys}}(\Gamma)\|_\infty < 1
  1. Variational principle (Theorem 2.1 [T]): dynamics of θij\theta_{ij} follow from stationarity of the action δSGap=0\delta S_{\text{Gap}} = 0
  2. Action aa enters through hext(a)h^{\text{ext}}(a) → modifies the equation of motion for θij\theta_{ij}:
mijθ¨ij=VGapθij+κ(θijtargetθij)Γ2θ˙ij+hijext(a)m_{ij}\ddot{\theta}_{ij} = -\frac{\partial V_{\text{Gap}}}{\partial \theta_{ij}} + \kappa(\theta_{ij}^{\text{target}} - \theta_{ij}) - \Gamma_2 \dot{\theta}_{ij} + h^{\text{ext}}_{ij}(a)
  1. Minimisation of σsys\|\sigma_{\mathrm{sys}}\|_\infty is the unique criterion equivalent to maximising the distance to the boundary V\mathcal{V} (viability region) in the metric induced by σsys\sigma_{\mathrm{sys}}. \blacksquare

3.2 Motor stress (T-159)

Theorem T-159 (Profile-relative motor stress) [T]

Statement [T]

For a holon with self-model ρ=φ(Γ)\rho_* = \varphi(\Gamma), the motor stress is defined as:

σkmotor(Γ):=1γkkρkk,k=1,,7\sigma^{\mathrm{motor}}_k(\Gamma) := 1 - \frac{\gamma_{kk}}{\rho^*_{kk}}, \quad k = 1, \ldots, 7

Action selection — minimisation of the maximum deficit (signed maximum):

a=argminaAmaxkσkmotor(Γ(τ+δτa))a^* = \arg\min_{a \in \mathcal{A}} \max_k \sigma^{\mathrm{motor}}_k(\Gamma(\tau + \delta\tau \mid a))

maxk\max_k (signed) is used rather than maxk\max_k |\cdot| (sup-norm): a resource surplus (σkmotor<0\sigma^{\mathrm{motor}}_k < 0) is not penalised; only a deficit (σkmotor>0\sigma^{\mathrm{motor}}_k > 0) is penalised. This provides a directed signal: approaching a resource reduces the deficit, approaching danger increases it.

Proof.

Step 1 (Equilibrium). σkmotor=0    γkk=ρkk\sigma^{\mathrm{motor}}_k = 0 \iff \gamma_{kk} = \rho^*_{kk}. At the attractor ρΩ\rho^*_\Omega, where R[Γ]=κ(ρΓ)gV=0\mathcal{R}[\Gamma] = \kappa(\rho_* - \Gamma) \cdot g_V = 0 (balance), γkk=ρkk\gamma_{kk} = \rho^*_{kk} and motor stress vanishes — the system is "satisfied".

Step 2 (Sign and gradient). σkmotor/γkk=1/ρkk<0\partial\sigma^{\mathrm{motor}}_k / \partial\gamma_{kk} = -1/\rho^*_{kk} < 0. Increasing γkk\gamma_{kk} (resource growth in channel kk) decreases motor stress. This is consistent with regeneration R=κ(ρΓ)\mathcal{R} = \kappa(\rho_* - \Gamma), which pulls γkk\gamma_{kk} toward ρkk\rho^*_{kk}, reducing σkmotor|\sigma^{\mathrm{motor}}_k|.

Step 3 (Sensitivity of critical channels). σkmotor/γkk=1/ρkk|\partial\sigma^{\mathrm{motor}}_k / \partial\gamma_{kk}| = 1/\rho^*_{kk}. For small ρkk\rho^*_{kk} (critical sectors A, S, D with ρkk0.05\rho^*_{kk} \approx 0.05) sensitivity 20\approx 20; for large ones (E, O, U with ρkk0.25\rho^*_{kk} \approx 0.25) — 4\approx 4. Small channels react more sharply — correct prioritisation of survival.

Step 4 (Convergence to T-92 at the boundary). As PPcrit=2/7P \to P_{\mathrm{crit}} = 2/7 the self-model φ(Γ)I/7\varphi(\Gamma) \to I/7 (canonical Fano-channel target at P=2/7P = 2/7, T-126). Then ρkk1/7\rho^*_{kk} \to 1/7 and:

σkmotor=1γkk1/7=17γkk=σk(canonical T-92 [T])\sigma^{\mathrm{motor}}_k = 1 - \frac{\gamma_{kk}}{1/7} = 1 - 7\gamma_{kk} = \sigma_k \quad \text{(canonical T-92 [T])}

Step 5 (G2G_2-invariance). γkk\gamma_{kk} and ρkk\rho^*_{kk} transform covariantly under G2G_2 (T-42a [T]). Their ratio is a G2G_2-invariant observable. \blacksquare

Relation to canonical σ_sys
  • T-92 / T-158 [T] define σsys\sigma_{\mathrm{sys}} with clamp[0,1][0,1] — a measure of viability (distance to V\partial\mathcal{V}). Used for DIAGNOSTICS.
  • T-159 [T] defines σmotor\sigma^{\mathrm{motor}} without clamp — a measure of motor deficit (distance to ρ\rho_*). Used for ACTION SELECTION.

When ρ=I/7\rho_* = I/7 (viability boundary) both coincide. When ρI/7\rho_* \neq I/7 (normal mode) motor stress provides a directed signal, while canonical σsys\sigma_{\mathrm{sys}} with clamp[0,1][0,1] loses information about channels with γkk>1/7\gamma_{kk} > 1/7.

3.3 Functor Dec

The action (decoding) functor:

Dec:(Γ,σmotor)aA\mathrm{Dec}: (\Gamma, \sigma^{\mathrm{motor}}) \mapsto a^* \in \mathcal{A}

Properties:

  • D-dimension as the primary motor channel: action is implemented through modification of h(D)h^{(D)} — the dynamic dimension DD controls the holon's "motor system"
  • σ-gradient descent: the practical algorithm — descent along amaxkσkmotor\nabla_a \max_k \sigma^{\mathrm{motor}}_k with the Fisher metric on D(C7)\mathcal{D}(\mathbb{C}^7):
at+1=atηF1(Γ)amaxkσkmotor(Γ(τ+δτat))a_{t+1} = a_t - \eta \cdot F^{-1}(\Gamma) \cdot \nabla_a \max_k \sigma^{\mathrm{motor}}_k(\Gamma(\tau + \delta\tau \mid a_t))

where F(Γ)F(\Gamma) is the Fisher information on D(C7)\mathcal{D}(\mathbb{C}^7).


4. Universal Encoder/Decoder Architecture

Perception → decision → action cycle:

StageMappingFormalismTheorem
PerceptionEnv → hexth^{\text{ext}}δΓ\delta\GammaEnc (CPTP)T-100 [T]
EvaluationΓ\Gammaσmotor\sigma^{\mathrm{motor}}1γkk/ρkk1 - \gamma_{kk}/\rho^*_{kk}T-159 [T]
Decisionσmotor\sigma^{\mathrm{motor}}aa^*argminamaxkσkmotor\arg\min_a \max_k \sigma^{\mathrm{motor}}_kT-159 [T]
Actionaa^*hext(a)h^{\text{ext}}(a^*) → EnvDecT-102 [T]
UpdateΓ\Gammaφ(Γ)\varphi(\Gamma)R\mathcal{R}Self-modellingT-62 [T]

5. Hedonic Mechanism

5.0 Intuition: why does a system need to "feel"

Why do living beings have pain and pleasure? The standard evolutionary biology answer: "to survive". But CC gives a more precise answer: hedonic valence is the derivative of viability with respect to the regenerative channel. Pleasure is not a "reward for correct behaviour" (as in RL), but a direct signal that the system is approaching its target state ρ\rho_*.

The key difference from reinforcement learning: in RL, reward is an external signal set by the designer. In CC, hedonic valence is an intrinsic property of the dynamics, derived from the evolution equation. Nobody "rewards" an amoeba for finding glucose — the change in dP/dτRdP/d\tau|_{\mathcal{R}} arises automatically when Γ\Gamma shifts toward ρ\rho_*.

Analogy: imagine a plant turning toward light. There is no "reward centre" telling the stem: "good, continue". There is a physicochemical process (auxin redistributes) that is simultaneously the movement and the "evaluation" — light amplifies the processes leading to growth. In CC, Vhed\mathcal{V}_{\text{hed}} plays an analogous role, but at the level of the coherence matrix.

5.1 Hedonic valence

Theorem T-103 (Hedonic valence) [T] + [I]

Statement

Hedonic valence is defined as the derivative of purity with respect to the regenerative channel:

Vhed:=dPdτR\mathcal{V}_{\text{hed}} := \left.\frac{dP}{d\tau}\right|_{\mathcal{R}}

where R|_{\mathcal{R}} denotes the contribution from the regenerative term R[Γ,E]\mathcal{R}[\Gamma, E] only.

Explanation. From the evolution equation:

dPdτ=2Tr(ΓDΩ[Γ])0, dissipation+2Tr(ΓR[Γ,E])Vhed\frac{dP}{d\tau} = \underbrace{-2\mathrm{Tr}(\Gamma \cdot \mathcal{D}_\Omega[\Gamma])}_{\leq 0,\text{ dissipation}} + \underbrace{2\mathrm{Tr}(\Gamma \cdot \mathcal{R}[\Gamma, E])}_{\mathcal{V}_{\text{hed}}}

(The Hamiltonian term does not change PP: Tr(Γ[H,Γ])=0\mathrm{Tr}(\Gamma [H, \Gamma]) = 0.)

Properties of valence:

PropertyFormulaInterpretation
PositiveVhed>0\mathcal{V}_{\text{hed}} > 0Γ\Gamma approaches ρ\rho_* → "pleasure"
NegativeVhed<0\mathcal{V}_{\text{hed}} < 0Γ\Gamma moves away from ρ\rho_* → "suffering"
ZeroVhed=0\mathcal{V}_{\text{hed}} = 0Balance or Γ=ρ\Gamma = \rho_* → "neutrality"

Epistemic stratification of T-103

T-103 contains three epistemic levels:

  1. Formula [T]: Vhed=2κ(Γ)gV(P)Tr(Γ(ρΓ))\mathcal{V}_{\text{hed}} = 2\kappa(\Gamma) \cdot g_V(P) \cdot \mathrm{Tr}(\Gamma \cdot (\rho_* - \Gamma)) — an identity from the evolution equation (substituting R=κ(ρΓ)gV(P)\mathcal{R} = \kappa(\rho_* - \Gamma) \cdot g_V(P)). An unconditional mathematical fact.

  2. Observability [T]: At L2 reflection level (R1/3R \geq 1/3) the replacement channel T-77 provides access to dP/dτRdP/d\tau|_{\mathcal{R}}. Thus, Vhed\mathcal{V}_{\text{hed}} is observable for any system with RRthR \geq R_{\mathrm{th}} — this is a consequence of T-77 [T], requiring no additional assumptions.

  3. Phenomenal interpretation [I]: Identification of Vhed>0\mathcal{V}_{\text{hed}} > 0 with "pleasure" and Vhed<0\mathcal{V}_{\text{hed}} < 0 with "suffering" — a semantic bridge between mathematics and phenomenology.

Life analogy: pleasure from hot tea

Imagine: you are cold and drinking hot tea. The first sip — delight (Vhed>0\mathcal{V}_{\text{hed}} > 0). The second — slightly weaker. By the fifth cup — neutrality (Vhed0\mathcal{V}_{\text{hed}} \approx 0). The sixth cup causes discomfort (Vhed<0\mathcal{V}_{\text{hed}} < 0) — you have "overheated".

What happened? Γ\Gamma (your state) was moving toward ρ\rho_* (the target — "warmed up"). As it approached, Tr(Γρ)P\mathrm{Tr}(\Gamma \cdot \rho_*) - P diminished, valence tended to zero. When Γ\Gamma "overshot" ρ\rho_* (overheating), overlap falls below PP, and Vhed\mathcal{V}_{\text{hed}} becomes negative. Nobody "programmed" you to stop drinking — the T-103 formula automatically generates the signal "enough".

The key difference from reinforcement learning: in RL the designer must specify a reward function (e.g., r=+1r = +1 for tea, 1-1 for overheating). In CC reward is derived from dynamics — Vhed\mathcal{V}_{\text{hed}} "knows" when to stop on its own, because it is nothing other than the rate of approach to the target state.

5.2 Relation to the target state

Substituting the canonical form R[Γ,E]=κ(Γ)(ρΓ)gV(P)\mathcal{R}[\Gamma, E] = \kappa(\Gamma) \cdot (\rho_* - \Gamma) \cdot g_V(P) [T]:

Vhed=2κ(Γ)gV(P)Tr(Γ(ρΓ))\mathcal{V}_{\text{hed}} = 2\kappa(\Gamma) \cdot g_V(P) \cdot \mathrm{Tr}(\Gamma \cdot (\rho_* - \Gamma))

When gV(P)=1g_V(P) = 1 (sufficient purity PPoptP \geq P_{\text{opt}}):

Vhed=2κ(Γ)[Tr(Γρ)P]\mathcal{V}_{\text{hed}} = 2\kappa(\Gamma) \cdot \left[\mathrm{Tr}(\Gamma \cdot \rho_*) - P\right]

The sign is determined by the ratio of overlap Tr(Γρ)\mathrm{Tr}(\Gamma \cdot \rho_*) to purity P=Tr(Γ2)P = \mathrm{Tr}(\Gamma^2):

  • If Γ\Gamma is far from ρ\rho_* and Tr(Γρ)>P\mathrm{Tr}(\Gamma \cdot \rho_*) > P, valence is positive — regeneration "pulls" toward ρ\rho_*
  • If Γρ\Gamma \approx \rho_*, then Tr(Γρ)P\mathrm{Tr}(\Gamma \cdot \rho_*) \approx P → valence tends to zero

6. 21 Qualia-types as Sensorimotor Channels

Each of the 21 off-diagonal coherences γij\gamma_{ij} (iji \neq j) represents a sensorimotor channel with a specific function:

6.1 Perceptual channels (perception)

ChannelCoherenceSensory roleFormal action
ApperceptionγAE\gamma_{AE}Conscious perceptionhAE(H)h^{(H)}_{AE}: articulation of input into the field of interiority
ActualisationγAD\gamma_{AD}Embodiment of perception in dynamicshAD(H)h^{(H)}_{AD}: transformation of input signal into action
RepresentationγSE\gamma_{SE}Structuring of experiencehSE(H)h^{(H)}_{SE}: creation of internal model
InductionγSL\gamma_{SL}Logical processing of structurehSL(H)h^{(H)}_{SL}: inference of patterns from data
GroundingγAO\gamma_{AO}Anchoring perception in groundhAO(D)h^{(D)}_{AO}: stabilisation of perception by memory
Experiential groundγEO\gamma_{EO}Rootedness of the subjectivehEO(R)h^{(R)}_{EO}: regeneration from deep experience
ContextγSO\gamma_{SO}Structure-in-contexthSO(D)h^{(D)}_{SO}: noise-robustness of patterns

6.2 Motor channels (action)

ChannelCoherenceMotor roleFormal action
RegulationγDL\gamma_{DL}Logical control of dynamicshDL(D)h^{(D)}_{DL}: management of computational process
TeleologyγDU\gamma_{DU}Goal-directedness of actionhDU(D)h^{(D)}_{DU}: alignment of dynamics with goals
AffectγDE\gamma_{DE}Emotional colouring of actionhDE(D)h^{(D)}_{DE}: modulation of dynamics by interiority
Action integrationγAU\gamma_{AU}Unity of the motor acthAU(H)h^{(H)}_{AU}: coordination of subsystems
Volitional effortγLU\gamma_{LU}Logically directed integrationhLU(R)h^{(R)}_{LU}: restoration of decision coherence
Action memoryγDO\gamma_{DO}Motor memoryhDO(D)h^{(D)}_{DO}: stabilisation of skills

6.3 Integrative channels

ChannelCoherenceIntegrative roleFormal action
InsightγLE\gamma_{LE}Logic-in-experiencehLE(R)h^{(R)}_{LE}: understanding as regeneration
NarrativeγAL\gamma_{AL}Articulation of logichAL(H)h^{(H)}_{AL}: shaping of reasoning
Grounded unityγOU\gamma_{OU}Ground of integrationhOU(R)h^{(R)}_{OU}: foundation of wholeness
Embodied unityγSU\gamma_{SU}Structure of integrationhSU(R)h^{(R)}_{SU}: architecture of connectivity
Living experienceγEU\gamma_{EU}Unity of experiencehEU(R)h^{(R)}_{EU}: integration as recovery
Dynamic groundγAS\gamma_{AS}Articulation of structurehAS(H)h^{(H)}_{AS}: external expression of inner order
Logical groundγLO\gamma_{LO}Logic-in-groundhLO(H)h^{(H)}_{LO}: formalisation of knowledge
Interpretation [I]

The division of 21 channels into perceptual, motor, and integrative is not strict: each γij\gamma_{ij} is simultaneously a sensory and a motor channel (through hijexth^{\text{ext}}_{ij}). The classification above reflects the dominant function — which of the three channels (h(H),h(D),h(R)h^{(H)}, h^{(D)}, h^{(R)}) is most active for that coherence.


7. Factorisation of Enc through Arbitrary Representations

7.1 Ontological projection

Corollary T-100a (Enc factorisation) [T]

Statement

For an arbitrary observation space ObsSpaceRD\mathrm{ObsSpace} \subseteq \mathbb{R}^D (DD — arbitrary dimension), the encoding functor T-100 factorises as:

Enc=πΓEncrepr\mathrm{Enc} = \pi_\Gamma \circ \mathrm{Enc}_{\text{repr}}

where:

  • Encrepr:ObsSpaceSRd\mathrm{Enc}_{\text{repr}}: \mathrm{ObsSpace} \to \mathcal{S} \subseteq \mathbb{R}^d — an arbitrary representation (feature map)
  • πΓ:SEnd(D(C7))\pi_\Gamma: \mathcal{S} \to \mathrm{End}(\mathcal{D}(\mathbb{C}^7)) — the ontological projection, unique up to G2G_2-gauge

Proof.

  1. By T-100 [T], Enc:ObsSpaceEnd(D(C7))\mathrm{Enc}: \mathrm{ObsSpace} \to \mathrm{End}(\mathcal{D}(\mathbb{C}^7)) is a CPTP functor.
  2. Any intermediate representation Encrepr:ObsSpaceS\mathrm{Enc}_{\text{repr}}: \mathrm{ObsSpace} \to \mathcal{S} defines a factorisation through πΓ=EncEncrepr1Im(Encrepr)\pi_\Gamma = \mathrm{Enc} \circ \mathrm{Enc}_{\text{repr}}^{-1}\big|_{\mathrm{Im}(\mathrm{Enc}_{\text{repr}})}.
  3. By T-102 [T], πΓ\pi_\Gamma decomposes into 3 channels: πΓ(s)=h(H)(s)h(D)(s)h(R)(s)\pi_\Gamma(s) = h^{(H)}(s) \oplus h^{(D)}(s) \oplus h^{(R)}(s).
  4. Uniqueness of πΓ\pi_\Gamma (up to G2G_2) — consequence of the uniqueness theorem [T]: constraints (AP)+(PH)+(QG)+(V) on D(C7)\mathcal{D}(\mathbb{C}^7) fix the projection. \blacksquare

Intuitively: the factorisation of Enc means that it does not matter how exactly you extract features from the input data. One can use a convolutional neural network, wavelet transform, or hand-crafted heuristics — this is Encrepr\mathrm{Enc}_{\text{repr}}, the arbitrary part. But the final step — the projection πΓ\pi_\Gamma from feature space into the space of Γ\Gamma-modifications — is unique. This is like saying: the route to the airport can be anything, but the runway is one.

For robotics this means: sensors can be arbitrary (camera, lidar, tactile array), preprocessing — anything, but the "last mile" of perception — the ontological projection πΓ\pi_\Gamma — is given by mathematics, not engineering choice.

7.2 Ontological bottleneck

Regardless of the input data dimensionality DD, all information is compressed into the 7×77 \times 7 coherence matrix Γ\Gamma with 48\leq 48 real parameters:

CharacteristicValue
Input dimensionalityDD — arbitrary (from D=1D = 1 to D106D \gg 10^6)
Intermediate representationdd — arbitrary
Output dimensionalitydimRD(C7)=48\dim_{\mathbb{R}} \mathcal{D}(\mathbb{C}^7) = 48 (fixed)
Information per steplog272.81\leq \log_2 7 \approx 2.81 bits (T-107 [T])

Corollary: Modality-agnosticism is a theorem, not a design choice. Formally: πΓ\pi_\Gamma does not depend on DD or on the structure of ObsSpace\mathrm{ObsSpace} (topology, metric). If two different observation spaces ObsSpace1RD1\mathrm{ObsSpace}_1 \subseteq \mathbb{R}^{D_1} and ObsSpace2RD2\mathrm{ObsSpace}_2 \subseteq \mathbb{R}^{D_2} produce the same CPTP channels on D(C7)\mathcal{D}(\mathbb{C}^7), they are indistinguishable for the holon.

7.3 Canonical form of the projection

By T-102 [T], πΓ\pi_\Gamma is implemented through three channels — modifications of the Hamiltonian, dissipative, and regenerative dynamics respectively:

πΓ(s)=(δH(s),  δD(s),  δR(s))End(D(C7))\pi_\Gamma(s) = \bigl(\delta H(s),\; \delta\mathcal{D}(s),\; \delta\mathcal{R}(s)\bigr) \in \mathrm{End}(\mathcal{D}(\mathbb{C}^7))

Practically this means that any implementation of Enc\mathrm{Enc} (from a simple sensor to a complex encoder) must end with the same 3-channel interface:

sSπΓ(hij(H)(s),  hij(D)(s),  hij(R)(s))R21R21R21s \in \mathcal{S} \xrightarrow{\pi_\Gamma} \bigl(h^{(H)}_{ij}(s),\; h^{(D)}_{ij}(s),\; h^{(R)}_{ij}(s)\bigr) \in \mathbb{R}^{21} \oplus \mathbb{R}^{21} \oplus \mathbb{R}^{21}

This structure is invariant: it is defined by G2G_2-symmetry and does not depend on the choice of representation Encrepr\mathrm{Enc}_{\text{repr}}.


8. Relation to Other Results

ResultRelationReference
T-57 (LGKS)Grounds T-102 (3-term completeness)Lindblad operators
T-62 (φ\varphi-operator)ρ\rho_* in the regeneration cycleSelf-observation
T-92 (σsys\sigma_{\mathrm{sys}})Optimality criterion in DecTheorem 10.1
T-75 (Schwinger–Keldysh)Lagrangian formulation with dissipationLagrangian
T-96 (Attractor)Non-trivial ρ\rho_* for guidanceEvolution
FEP (Theorem 4.1)Macroscopic limit of DecVariational principles
T-109–T-112 (Learning bounds)Lower bounds on learning rate through Enc/Dec cycleLearning bounds
T-113 (Minimality N=7)N=7 — minimal architecture for learningLearning bounds

9. Information Capacity of Enc (T-107) [T]

Theorem T-107 (Information capacity of Enc) [T]

Maximum information extractable by functor Enc per single observation:

CEncmax{po}χ({po,Enc(o)})log272.81 bits/observationC_{\mathrm{Enc}} \leq \max_{\{p_o\}} \chi(\{p_o, \mathrm{Enc}(o)\}) \leq \log_2 7 \approx 2.81 \text{ bits/observation}

where χ\chi is the Holevo quantity.

Proof.

  1. Functor Enc:ObsSpaceEnd(D(C7))\mathrm{Enc}: \mathrm{ObsSpace} \to \mathrm{End}(\mathcal{D}(\mathbb{C}^7)) maps observations to CPTP channels on D(C7)\mathcal{D}(\mathbb{C}^7) (T-100 [T]).
  2. The Holevo quantity is bounded by the output space dimensionality: χlog2dimH=log27\chi \leq \log_2 \dim \mathcal{H} = \log_2 7.
  3. From T-102 [T]: Enc(o)\mathrm{Enc}(o) decomposes into 3 channels, each acting on D(C7)\mathcal{D}(\mathbb{C}^7).
  4. The composite channel does not increase capacity (Holevo subadditivity):
CEncS(Γˉ)opoS(Enc(o)[Γ])Smax(D(C7))=log27C_{\mathrm{Enc}} \leq S(\bar{\Gamma}) - \sum_o p_o S(\mathrm{Enc}(o)[\Gamma]) \leq S_{\max}(\mathcal{D}(\mathbb{C}^7)) = \log_2 7

The upper bound is achieved for an ensemble of orthogonal pure states. \blacksquare

Corollary (Bounded rationality): The bound 2.81\leq 2.81 bits/observation is a derived bound, not a postulated one. Connection with Simon's bounded rationality: bounded rationality is not an empirical fact, but a consequence of N=7.


10. Compositionality of Enc/Dec (T-108) [T]

Theorem T-108 (Compositionality of Enc/Dec) [T]

For a composite of two holons, encoding preserves structure:

Enc12=Φagg(Enc1Enc2)\mathrm{Enc}_{12} = \Phi_{\mathrm{agg}} \circ (\mathrm{Enc}_1 \otimes \mathrm{Enc}_2)

where Φagg:D(C72)D(C7)\Phi_{\mathrm{agg}}: \mathcal{D}(\mathbb{C}^{7^2}) \to \mathcal{D}(\mathbb{C}^7) — CPTP aggregation from T-72 (CC-6) [T].

Proof.

  1. Enc1,Enc2\mathrm{Enc}_1, \mathrm{Enc}_2 — CPTP functors (T-100 [T]).
  2. Tensor product Enc1Enc2\mathrm{Enc}_1 \otimes \mathrm{Enc}_2 — a CPTP channel on D(C49)\mathcal{D}(\mathbb{C}^{49}).
  3. Aggregation Φagg\Phi_{\mathrm{agg}} — CPTP from Morita equivalence (T-58 [T]): D(C49)D(C7)\mathcal{D}(\mathbb{C}^{49}) \to \mathcal{D}(\mathbb{C}^7).
  4. Composition of CPTP channels — CPTP. Functoriality (Enc(o1o2)=Enc(o1)Enc(o2)\mathrm{Enc}(o_1 \circ o_2) = \mathrm{Enc}(o_1) \circ \mathrm{Enc}(o_2)) from T-100 is preserved under aggregation.
  5. Uniqueness — from G2G_2-rigidity at each scale (T-72 [T]). \blacksquare

Corollary for cognitive engineers: diagnostics (σ_sys, Enc/Dec monitoring) are the same at all scales — from individual agent to organisation.

Analogously for Dec:

Dec12=(Dec1Dec2)Φsplit\mathrm{Dec}_{12} = (\mathrm{Dec}_1 \otimes \mathrm{Dec}_2) \circ \Phi_{\mathrm{split}}

where Φsplit\Phi_{\mathrm{split}} — the inverse map (splitting the composite σ into components).


11. Temporal Integration

11.1 Cumulative capacity

Corollary T-107a (Cumulative information) [T]

Statement

Over nn successive observations a holon accumulates information about the environment:

Innlog272.81n  bitsI_n \leq n \cdot \log_2 7 \approx 2.81\,n \;\text{bits}

The upper bound is achievable when successive observations are informationally independent.

Proof.

  1. By T-107 [T], one observation brings log27\leq \log_2 7 bits.
  2. Holevo subadditivity: χ({po1,,on})k=1nχ({pok})\chi(\{p_{o_1,\ldots,o_n}\}) \leq \sum_{k=1}^n \chi(\{p_{o_k}\}).
  3. For independent observations the inequality becomes an equality. \blacksquare

11.2 Minimum number of observations

Corollary T-107b (Minimum observations) [T]

Statement

For an environment with information entropy IenvI_{\mathrm{env}} bits, the minimum number of observations for complete encoding:

nmin=Ienvlog27n_{\min} = \left\lceil\frac{I_{\mathrm{env}}}{\log_2 7}\right\rceil

Proof. Direct consequence of T-107a: In2.81nI_n \leq 2.81\,n, hence nIenv/log27n \geq I_{\mathrm{env}} / \log_2 7. \blacksquare

Corollary for complex modalities. Encoding an environment with high information complexity (large IenvI_{\mathrm{env}}) inevitably requires a multi-step process. This is not an implementation limitation, but a fundamental bound following from dimH=7\dim \mathcal{H} = 7.

Relation to T-109 (information learning bound): T-107b gives a lower bound on perception, T-109 gives a lower bound on learning (including stabilisation of the solution). Always noptnminn_{\mathrm{opt}} \geq n_{\min}, since learning includes perception as a subtask. See learning bounds.

11.3 Information absorption rate

Define the information absorption rate:

I˙(τ)=dIdτ=χ({po,Enc(o)[Γ(τ)]})\dot{I}(\tau) = \frac{dI}{d\tau} = \chi\bigl(\{p_o,\, \mathrm{Enc}(o)[\Gamma(\tau)]\}\bigr)

From T-107 [T]: I˙(τ)log27\dot{I}(\tau) \leq \log_2 7 for any τ\tau.

The actual rate depends on the current state Γ(τ)\Gamma(\tau):

  • When ΓI/7\Gamma \approx I/7 (maximally mixed): I˙0\dot{I} \to 0 — system is "deafened", distinguishability is minimal
  • When P2/7P \gg 2/7 (high purity): I˙log27\dot{I} \to \log_2 7 — maximum distinguishability
  • When P<2/7P < 2/7 (non-viability): encoding degrades, T-104 is not satisfied

12. Predictive Structure of Enc

12.1 Optimal Enc as a ΔF maximiser

Corollary T-107c (Predictive optimality of Enc) [T]

Statement

The optimal encoding functor Enc\mathrm{Enc}^* maximises available free energy:

Enc=argmaxEncΔF(Enc(o)[Γ],ρ)\mathrm{Enc}^* = \arg\max_{\mathrm{Enc}} \Delta F\bigl(\mathrm{Enc}(o)[\Gamma],\, \rho_*\bigr)

where ΔF=Tr(R[Γ,E](ρΓ))\Delta F = \mathrm{Tr}\bigl(\mathcal{R}[\Gamma, E] \cdot (\rho_* - \Gamma)\bigr)free energy.

Proof.

  1. By the variational principle (Theorem 4.1 [T]): stationary dynamics of Γ\Gamma minimise Friston's free energy F[Γ]=KL(Γρ)+H[Γ]F[\Gamma] = \mathrm{KL}(\Gamma \| \rho_*) + H[\Gamma].
  2. Functor Enc(o)\mathrm{Enc}(o) modifies ΓΓ\Gamma \to \Gamma'. The optimal modification is the one that maximally increases ΔF=F[Γ]F[Γ]\Delta F = F[\Gamma] - F[\Gamma'].
  3. Maximisation of ΔF\Delta F is equivalent to maximising KL(Γρ)-\mathrm{KL}(\Gamma' \| \rho_*) at fixed entropy — i.e., approaching the target state.
  4. From T-107 [T]: ΔFCEnclog27\Delta F \leq C_{\mathrm{Enc}} \leq \log_2 7 per step — the upper bound is saturated. \blacksquare

12.2 Prediction error through 3 channels

Prediction error (discrepancy between expected and actual observation) decomposes across three channels (T-102 [T]):

δpred=Enc(oreal)Enc(opred)=(δh(H))2+(δh(D))2+(δh(R))2\delta_{\mathrm{pred}} = \bigl\|\mathrm{Enc}(o_{\mathrm{real}}) - \mathrm{Enc}(o_{\mathrm{pred}})\bigr\| = \sqrt{(\delta h^{(H)})^2 + (\delta h^{(D)})^2 + (\delta h^{(R)})^2}

Each channel contributes a specific type of error:

ChannelErrorInterpretation
δh(H)\delta h^{(H)}EnergeticUnexpected structure of environment
δh(D)\delta h^{(D)}NoiseUnexpected level of stochasticity
δh(R)\delta h^{(R)}RegenerativeUnexpected change in target state

Relation to the hedonic mechanism: By T-103 [T]+[I], an error in the regenerative channel (δh(R)0\delta h^{(R)} \neq 0) directly modulates Vhed\mathcal{V}_{\mathrm{hed}} — unexpected influences on regeneration are experienced as a change in valence.


13. Multimodal Decomposition

13.1 Composition of modalities

Corollary T-108a (Multimodal decomposition) [T]

Statement

For MM independent perceptual modalities with functors Encm:ObsSpacemEnd(D(C7))\mathrm{Enc}_m: \mathrm{ObsSpace}_m \to \mathrm{End}(\mathcal{D}(\mathbb{C}^7)), joint encoding:

Enc(o1,,oM)=m=1MwmEncm(om)+m<mΔmm\mathrm{Enc}(o_1, \ldots, o_M) = \sum_{m=1}^{M} w_m \cdot \mathrm{Enc}_m(o_m) + \sum_{m < m'} \Delta_{mm'}

where wm0w_m \geq 0, wm=1\sum w_m = 1 — modality weights, Δmm\Delta_{mm'} — cross-modal coupling.

Proof.

  1. By T-100 [T], each Encm\mathrm{Enc}_m is a CPTP functor.
  2. Convex combination of CPTP channels — CPTP: wmEncm\sum w_m \mathrm{Enc}_m is defined when wm=1\sum w_m = 1.
  3. Cross-modal terms Δmm\Delta_{mm'} — CPTP corrections of order O(γij)O(|\gamma_{ij}|), where γij\gamma_{ij} are coherences linking dimensions engaged by modalities mm and mm'.
  4. From T-108 [T] (compositionality): aggregation of modalities preserves CPTP structure and functoriality. \blacksquare

13.2 Competition for capacity

From T-107 [T], total capacity of MM modalities per step:

m=1MwmCmlog27\sum_{m=1}^{M} w_m \cdot C_m \leq \log_2 7

Corollary: MM modalities compete for the fixed bandwidth of 2.812.81 bits/step. Increasing the number of modalities MM at fixed nn does not increase total information — it merely distributes it among channels.

13.3 Attention as optimal allocation

Optimal weights wmw_m^* are determined by maximising ΔF\Delta F:

wm=ΔFmmΔFmw_m^* = \frac{\Delta F_m}{\sum_{m'} \Delta F_{m'}}

where ΔFm=ΔF(Encm(om)[Γ],ρ)\Delta F_m = \Delta F\bigl(\mathrm{Enc}_m(o_m)[\Gamma],\, \rho_*\bigr) — contribution of modality mm to free energy.

Interpretation [I]: The optimal allocation of weights wmw_m^* formally coincides with the structure of attention — encoding resources are directed where the informational value (ΔFm\Delta F_m) is maximal. This is not an additional postulate: attention is a consequence of Enc optimality under bounded capacity (T-107).

13.4 Cross-modal coupling

The terms Δmm\Delta_{mm'} are determined by coherences γij\gamma_{ij}, where ii and jj are dimensions engaged by different modalities:

Δmmγijmin(wm,wm)\|\Delta_{mm'}\| \leq |\gamma_{ij}| \cdot \min(w_m, w_{m'})

Corollary: Cross-modal integration is only possible when coherences between the corresponding dimensions are non-zero. Fully decohered dimensions (γij=0|\gamma_{ij}| = 0) do not admit multimodal binding — modalities remain isolated.


14. Comparison with Classical Approaches

The sensorimotor theory of CC did not arise in a vacuum — it answers questions posed by three powerful traditions: classical control theory, active inference, and reinforcement learning. In this section we conduct a systematic comparison, showing where CC coincides with each tradition and where it fundamentally diverges.

14.1 CC vs. classical control theory

Classical control theory (Wiener, Kalman, Pontryagin) describes the "sensor → controller → actuator" cycle through transfer functions, state space, and optimality criteria (LQR, H-infinity, etc.).

AspectClassical controlCC
State spaceRn\mathbb{R}^n, arbitrary nnD(C7)\mathcal{D}(\mathbb{C}^7), fixed
Optimality criterionQuadratic J=(xTQx+uTRu)dtJ = \int (x^T Q x + u^T R u)\,dtMin-max: minamaxkσkmotor\min_a \max_k \sigma^{\mathrm{motor}}_k
Number of control channelsArbitrary (design choice)Exactly 3 (Theorem T-102)
ObserverExternal (Kalman filter)Internal (φ(Γ)\varphi(\Gamma) — self-model)
ExperienceAbsentVhed\mathcal{V}_{\text{hed}} — hedonic valence
ScalingProblematic (curse of dimensionality)T-108: compositionality preserved

Key difference: A PID controller minimises a weighted sum of errors — and may allow catastrophe in one channel, compensating with success in another. CC uses the min-max strategy (T-159), which guarantees that no channel ends up in a critical state. This is not a heuristic but a consequence of viability being defined by the sup-norm of the stress tensor (T-92).

Where they coincide: In the linear approximation near ρ\rho_*, the evolution equation for Γ\Gamma reduces to a linear feedback system — standard control theory turns out to be a projection of CC onto the linear regime.

14.2 CC vs. active inference (FEP)

The Free Energy Principle (Friston, 2006) postulates that living systems minimise variational free energy F=KL(qp)+constF = \mathrm{KL}(q \| p) + \mathrm{const}, where qq is the internal model, pp is the generative model of the environment.

AspectActive inference (FEP)CC
Objective functionMinimise F=KL(qp)F = \mathrm{KL}(q \| p)Minimise maxkσkmotor\max_k \sigma^{\mathrm{motor}}_k
Generative modelPostulatedρ=φ(Γ)\rho_* = \varphi(\Gamma) — derived
Number of perception channelsUnboundedlog27\leq \log_2 7 bits/step (T-107)
ActionMinimise expected free energyargminamaxkσkmotor\arg\min_a \max_k \sigma^{\mathrm{motor}}_k
Subjective experienceNot explained$\mathcal{V}_{\text{hed}} = dP/d\tau
Ontological statusPrinciple (axiom)Consequence (Theorem 4.1 of CC)

Key difference: FEP is a principle: it postulates that systems minimise free energy, but does not explain where this principle comes from. In CC, minimisation of free energy is a theorem (Theorem 4.1 [T]): it is derived from the canonical evolution equation in the macroscopic limit. Moreover, CC shows that FEP is an approximation valid at P2/7P \gg 2/7; near PcritP_{\mathrm{crit}} corrections arise that FEP does not capture.

Where they coincide: The optimal Enc maximises ΔF\Delta F (T-107c [T]) — this is the exact analogue of "perceptual inference" in FEP. The functor Dec minimises σmotor\sigma^{\mathrm{motor}}, which in the macroscopic limit is equivalent to "active inference". Thus, FEP is a projection of CC sensorimotor theory onto the classical (non-quantum-coherent) regime.

14.3 CC vs. reinforcement learning (RL)

Reinforcement learning (Sutton, Barto) models an agent maximising cumulative reward Gt=k=0γkrt+kG_t = \sum_{k=0}^{\infty} \gamma^k r_{t+k} through interaction with the environment.

AspectRLCC
RewardExternal rtr_t (set by designer)Intrinsic Vhed\mathcal{V}_{\text{hed}} (derived)
Policyπ(as)\pi(a \mid s) — stochasticargminamaxkσkmotor\arg\min_a \max_k \sigma^{\mathrm{motor}}_k — deterministic
CriterionmaxE[γkrk]\max \mathbb{E}[\sum \gamma^k r_k]minmaxkσkmotor\min \max_k \sigma^{\mathrm{motor}}_k
Observation capacityUnbounded2.81\leq 2.81 bits/step (T-107)
Credit assignment problemTemporal difference, n-step, GAEImmediate: σkmotor\sigma^{\mathrm{motor}}_k — current deficit
ScalingProblematic (reward shaping, multi-agent)T-108: compositionality
Exploration vs. exploitationSeparate problemFollows from σ-gradient

Key difference: In RL reward is a "black box": the designer specifies rtr_t, and the agent maximises it. The credit assignment problem — which past actions led to the current reward — is one of the central challenges. In CC reward is not needed: motor stress σkmotor\sigma^{\mathrm{motor}}_k is an immediate, component-wise signal that tells which exact channel needs a resource and by how much. Credit assignment is solved automatically — through the 7-component structure of σ\sigma.

Where they coincide: If the 7-component σmotor\sigma^{\mathrm{motor}} is collapsed to a scalar (e.g., rt=maxkσkmotorr_t = -\max_k \sigma^{\mathrm{motor}}_k), Dec becomes formally equivalent to a greedy policy in RL with immediate reward. Thus, RL is a projection of CC sensorimotor theory onto scalar reward and stochastic policy.

14.4 Summary table

PropertyClassical controlFEPRLCC
Number of channelsArbitraryArbitrary1 (scalar rr)3 (theorem)
OntologyExternalGenerative modelMDPD(C7)\mathcal{D}(\mathbb{C}^7)
ExperienceNoNoNoVhed\mathcal{V}_{\text{hed}} [T]+[I]
ScalingDifficultLimitedDifficultT-108 [T]
AttentionSeparate modulePrecision weightingNoConsequence of T-107
StatusEngineeringPrincipleAlgorithmTheory

15. Worked Examples

To keep the formalism from remaining abstract, let us examine three examples of the sensorimotor cycle in operation — from simplest to complex.

15.1 Example 1: Bacterial chemotaxis

E. coli swims along a glucose gradient. Its sensorimotor cycle in CC terms:

Step 1 (Enc). Chemoreceptors on the membrane register concentration c(x)c(x). This modifies:

  • hAO(H)h^{(H)}_{AO}: articulation-ground (distinguishing "nutritious / not nutritious")
  • hDO(D)h^{(D)}_{DO}: dynamics-ground (environmental turbulence as noise)

Step 2 (σ-evaluation). The bacterium is "hungry" → γOO\gamma_{OO} is small → σOmotor=1γOO/ρOO>0\sigma^{\mathrm{motor}}_O = 1 - \gamma_{OO}/\rho^*_{OO} > 0. Channel O (ground) is in deficit.

Step 3 (Dec). maxkσkmotor=σOmotor\max_k \sigma^{\mathrm{motor}}_k = \sigma^{\mathrm{motor}}_O. Optimal action: move up the gradient c(x)c(x) → modification of hDO(D)h^{(D)}_{DO} via the flagellar motor.

Step 4 (Update). Glucose absorption → increase of γOO\gamma_{OO} → decrease of σOmotor\sigma^{\mathrm{motor}}_O. If a chemical stressor simultaneously arises, σDmotor\sigma^{\mathrm{motor}}_D may exceed σOmotor\sigma^{\mathrm{motor}}_O, and the bacterium switches to avoidance — the min-max strategy in action.

Interiority level: L0 (non-zero E-projection, but no self-observation). Vhed\mathcal{V}_{\text{hed}} is formally defined but not observable by the bacterium itself (R<1/3R < 1/3).

15.2 Example 2: Robotic manipulator

A robot assembles an object from a table. Its Γ\Gamma is initialised through quasi-functor GG from joint position data, camera image, and force-torque sensor.

Enc (multimodal):

  • Camera → Encvis\mathrm{Enc}_{\text{vis}}: hAS(H)h^{(H)}_{AS} (articulation of structure — object shape), hSE(H)h^{(H)}_{SE} (representation — internal scene model)
  • Proprioception → Encprop\mathrm{Enc}_{\text{prop}}: hDL(D)h^{(D)}_{DL} (regulation — current configuration)
  • Force-torque sensor → Encforce\mathrm{Enc}_{\text{force}}: hDO(D)h^{(D)}_{DO} (motor memory — contact forces)

Attention weights by T-108a: wvisΔFvisw^*_{\text{vis}} \propto \Delta F_{\text{vis}}. If the object is visible but not yet grasped — ΔFvis\Delta F_{\text{vis}} is large (need to refine the model). After grasping — ΔFforce\Delta F_{\text{force}} grows (need to control force), and attention automatically switches to the force-torque sensor.

Dec: σDmotor>0\sigma^{\mathrm{motor}}_D > 0 (dynamic deficit: arm not in required position) → action: move manipulator. As it approaches σDmotor0\sigma^{\mathrm{motor}}_D \to 0, and σLmotor>0\sigma^{\mathrm{motor}}_L > 0 may emerge (logical deficit: grasp plan not yet formed) → switch to planning.

15.3 Example 3: Human in an unfamiliar city

A person searches for a café. All 7 channels are active:

Channelσkmotor\sigma^{\mathrm{motor}}_kInterpretation
AA0.1Distinguishes signs — weak deficit
SS0.3No map of the area — moderate deficit
DD0.0Physically mobile — no deficit
LL0.2Route logic is incomplete
EE-0.1Curiosity (excess of interiority)
OO0.6Hungry — maximum deficit
UU0.1Internally composed

maxkσkmotor=σOmotor=0.6\max_k \sigma^{\mathrm{motor}}_k = \sigma^{\mathrm{motor}}_O = 0.6. Action is directed toward reducing deficit O: walk toward the nearest café. Along the way σSmotor\sigma^{\mathrm{motor}}_S may grow (got lost), and if σSmotor>σOmotor\sigma^{\mathrm{motor}}_S > \sigma^{\mathrm{motor}}_O, the person switches to orientation — stops, takes out phone, opens map.

Hedonic valence: Approaching the café increases γOO\gamma_{OO}Vhed>0\mathcal{V}_{\text{hed}} > 0 (anticipation). If the café is closed — sharp Vhed<0\mathcal{V}_{\text{hed}} < 0 (disappointment). This is not a metaphor: the formula Vhed=2κgVTr(Γ(ρΓ))\mathcal{V}_{\text{hed}} = 2\kappa \cdot g_V \cdot \mathrm{Tr}(\Gamma \cdot (\rho_* - \Gamma)) gives a quantitative prediction verifiable through physiological correlates (skin conductance, pupillometry).


Summary

  1. T-100 [T]: Encoding functor Enc exists and is unique (up to G2G_2)
  2. T-101 [T]: Viability diagnostic criterion = argminσsys\arg\min \|\sigma_{\mathrm{sys}}\|_\infty
  3. T-159 [T]: Motor stress σkmotor=1γkk/ρkk\sigma^{\mathrm{motor}}_k = 1 - \gamma_{kk}/\rho^*_{kk} — action selection through argminamaxkσkmotor\arg\min_a \max_k \sigma^{\mathrm{motor}}_k (signed max)
  4. T-102 [T]: The 3-term equation is complete — a fourth type of CPTP generator is impossible
  5. T-103 [T]+[I]: Hedonic valence = dP/dτRdP/d\tau|_{\mathcal{R}} (formula [T], interpretation [I])
  6. T-107 [T]: Information capacity log272.81\leq \log_2 7 \approx 2.81 bits/observation
  7. T-108 [T]: Enc/Dec preserved under composition (scale invariance of sensorimotor theory)
  8. Corollary T-100a [T]: Enc factorises through arbitrary representation → modality agnosticism
  9. Corollary T-107a/b [T]: Cumulative capacity In2.81nI_n \leq 2.81\,n bits → complex modalities require nmin=Ienv/log27n_{\min} = \lceil I_{\mathrm{env}} / \log_2 7 \rceil steps
  10. Corollary T-107c [T]: Optimal Enc maximises ΔF\Delta F (predictive structure)
  11. Corollary T-108a [T]: MM modalities compete for 2.812.81 bits/step → attention is optimal allocation

Conclusion

The sensorimotor theory of Coherence Cybernetics closes the formal cycle: environment → perception (Enc) → state (Γ\Gamma) → evaluation (σmotor\sigma^{\mathrm{motor}}) → action (Dec) → environment. All operations are implemented within the canonical 3-term evolution equation without additional postulates.

Let us summarise the three central achievements of this chapter:

First, we showed that interaction with the environment does not require expanding the evolution equation. Theorem T-102 [T] proves that any CPTP-compatible external influence decomposes into three channels — Hamiltonian, dissipative, and regenerative. A fourth type of influence is mathematically forbidden. This is a strong result: it means that the entire phenomenology of sensorimotor interaction — from bacterial chemotaxis to human navigation in a city — is described by the same 3-channel formalism.

Second, we derived internal "reward" from dynamics, rather than postulating it externally. Hedonic valence Vhed=dP/dτR\mathcal{V}_{\text{hed}} = dP/d\tau|_{\mathcal{R}} (T-103 [T]) is a mathematical identity, requiring neither a designer (as in RL) nor a principle (as in FEP). The phenomenal interpretation (Vhed>0\mathcal{V}_{\text{hed}} > 0 as "pleasure") remains at the [I]-level, but the formula itself is an unconditional theorem.

Third, we established fundamental limits on perception. Information capacity log272.81\leq \log_2 7 \approx 2.81 bits/observation (T-107 [T]) is not an empirical limitation or an engineering trade-off, but a consequence of dimH=7\dim \mathcal{H} = 7. Simon's bounded rationality, competition of modalities for attention, the necessity of multi-step perception of complex scenes — all of this is derived as consequences of a single theorem.

The theory is modality-agnostic: from the simplest sensors (D=1D = 1) to complex multimodal systems (D1D \gg 1) — the ontological projection πΓ\pi_\Gamma is unique and invariant. The factorisation Enc = πΓEncrepr\pi_\Gamma \circ \mathrm{Enc}_{\text{repr}} (T-100a [T]) separates "engineering freedom" (choice of representation) and "mathematical necessity" (projection into D(C7)\mathcal{D}(\mathbb{C}^7)).

Comparison with classical approaches (Section 14) showed that CC does not cancel but includes control theory, active inference, and reinforcement learning as special cases — projections of the full 7-dimensional coherent dynamics onto the linear, variational, and scalar-reward regimes respectively.

The next step — applying this formalism to stability problems and learning, where the sensorimotor cycle turns out to be not just a diagram but a concrete computational algorithm with provable bounds.


What we learned

  1. The environment does not add a 4th term (T-102 [T]): any external influence decomposes into Hamiltonian, dissipative, and regenerative channels. A fourth type is mathematically forbidden.
  2. Perception is not recording, but deformation of dynamics (T-100 [T]): the functor Enc maps an observation to a modification of the evolution equation, in a unique (up to G2G_2) way.
  3. Action is a min-max strategy (T-159 [T]): the system eliminates the largest deficit, not minimises "average error". No channel is left unattended.
  4. Pleasure and suffering are derivatives of viability (T-103 [T]+[I]): Vhed=dP/dτR\mathcal{V}_{\text{hed}} = dP/d\tau|_{\mathcal{R}} — a mathematical identity, requiring no external "reward designer".
  5. Fundamental bottleneck: log272.81\leq \log_2 7 \approx 2.81 bits/observation (T-107 [T]). Simon's bounded rationality is not an empirical fact but a consequence of N=7N = 7.
  6. Scale invariance (T-108 [T]): Enc/Dec preserved under composition. From bacterium to organisation — the same formal structure.
  7. Classical approaches are projections of CC: control theory, FEP, and RL are special cases — projections of the full 7-dimensional coherent dynamics.
Bridge to the next chapter

We have built the complete "perception-decision-action" cycle. But how robust is this cycle? What blow can it withstand? Where is the boundary between recoverable trauma and irreversible destruction? In the next chapter we will answer these questions: derive the stability radius formula rstab=P2/7r_{\mathrm{stab}} = \sqrt{P - 2/7}, trace the mechanism of the "death spiral" — and show that antifragility is not a metaphor, but a consequence of integration of experience.


Related documents: