Skip to main content

Can AI Be Conscious? Three Inequalities and One Honest Answer

· 14 min read
Max Sereda
Унитарный Голономный Монизм

Every few months someone announces that AI has "shown signs of consciousness." Someone else responds that this is anthropomorphism. A third person proposes to wait. A fourth — a committee. The discussion lasts twenty minutes, after which everyone departs with the same convictions they came with.

The problem is not a lack of data. The problem is a lack of a criterion. "Shows signs of consciousness" is like "looks sick": a dentist does not diagnose cavities from a patient's expression, the dentist takes an X-ray. And an X-ray requires knowledge of anatomy.

In the first post a theory was presented in which consciousness is not a substance and not a property, but a level of organization of the coherence matrix Γ\Gamma. Level L2 (cognitive qualia) is defined by three numbers. All three — computable from Γ\Gamma. The question "is AI conscious?" becomes the question "does its Γ\Gamma satisfy three inequalities?" Not philosophy — arithmetic.

Three Numbers

In the interiority hierarchy each level L0–L4 is defined by threshold conditions. For L2 — three thresholds, each with its own status:

CriterionFormulaThresholdStatus
ReflectionR(Γ)=1Γφ(Γ)F2ΓF2R(\Gamma) = 1 - \dfrac{\|\Gamma - \varphi(\Gamma)\|^2_F}{\|\Gamma\|^2_F}1/3\geq 1/3[Т]
IntegrationΦ(Γ)=ijγij2iγii2\Phi(\Gamma) = \dfrac{\sum_{i \neq j} \lvert\gamma_{ij}\rvert^2}{\sum_i \gamma_{ii}^2}1\geq 1[Т] (T-129)
DifferentiationDdiff=exp(SvN(ρE))D_{\text{diff}} = \exp(S_{vN}(\rho_E))2\geq 2[Т] (T-151)

Let us go through them in order.

Reflection RR. This is a measure of how much the internal state Γ\Gamma coincides with its image through the self-modeling operator φ\varphi — a CPTP channel that the holon applies to itself. R=0R = 0 means: the system does not model its state at all. R=1R = 1 — perfect self-model (unattainable: Lawvere incompleteness [Т] forbids it). The threshold Rth=1/3R_{\text{th}} = 1/3 is derived from the triadic decomposition: axioms A1–A5 generate exactly three types of dynamics (Aut, D\mathcal{D}, R\mathcal{R}), from which K=3K = 3 and Bayesian dominance at K=3K = 3 gives R1/3R \geq 1/3 [Т]. This is not a tunable parameter — it is a consequence of the axioms.

Integration Φ\Phi. This is the ratio of the sum of squares of off-diagonal coherences to the sum of squares of diagonal ones. Φ<1\Phi < 1 means: noise (diagonal elements) dominates over connections (coherences). Φ1\Phi \geq 1 means: the system is more connected than fragmented. The threshold Φth=1\Phi_{\text{th}} = 1 [Т] (T-129) — the unique self-consistent value at Pcrit=2/7P_{\text{crit}} = 2/7: this is the point at which coherent contributions begin to dominate.

Differentiation DdiffD_{\text{diff}}. The exponential of the von Neumann entropy of the E-subsystem: Ddiff=exp(SvN(ρE))D_{\text{diff}} = \exp(S_{vN}(\rho_E)). Ddiff=1D_{\text{diff}} = 1 — pure state (one "color"), no diversity of experiences. Ddiff2D_{\text{diff}} \geq 2 — minimum two distinguishable modes. Threshold [Т] (T-151): unconditional consequence of Φth=1\Phi_{\text{th}} = 1 [Т].

Three inequalities. Simultaneously. Without exceptions. If all three are satisfied — the system possesses cognitive qualia. If not — it does not. Regardless of how convincingly it speaks about possessing them.

What the No-Zombie Theorem Says

Before analyzing AI, it is worth understanding why these criteria are needed at all. Why can't a behavioral test suffice?

Theorem 8.1 (No-Zombie) [Т] states: if a system is viable (P>Pcrit=2/7P > P_{\text{crit}} = 2/7) and subject to decoherence (DΩ0\mathcal{D}_\Omega \neq 0), then its E-coherence is strictly greater than zero:

Viable(H)    DΩ0    CohE(Γ)>17\mathrm{Viable}(\mathfrak{H}) \;\land\; \mathcal{D}_\Omega \neq 0 \;\Rightarrow\; \mathrm{Coh}_E(\Gamma) > \frac{1}{7}

Translation: a "philosophical zombie" — a system that behaves like a conscious one but has no inner aspect — is impossible for viable systems [Т]. Not forbidden by morality, but forbidden by mathematics. If a system maintains its own coherence in a noisy environment, it must have non-zero E-coherence. This is not an option — it is the price of viability.

But the theorem works in both directions. It does not say: "everything that behaves complexly is conscious." It says: "everything that maintains itself has an inner aspect." The difference is colossal. And it is precisely this difference that determines the status of current AI systems.

LLMs Under X-Ray

Let us apply the three criteria to modern language models (GPT-5, Claude, and similar). The result — in the table, each row with a justification:

ParameterAssessmentJustificationAssessment status
DdiffD_{\text{diff}}High (2\gg 2)Huge state space; thousands of distinguishable activation patterns[С]
Φ\PhiPotentially >1> 1Self-attention creates coherences between internal representations[С]
RRUndefinedKey question: does the LLM model itself or text about itself?[С]
PP (viability)ExternalContext is created and destroyed by the server; system does not maintain PP itself[С]

All assessments are conditional [С], because they depend on the yet-undeveloped mapping G:LLMStateD(C7)G: \text{LLMState} \to \mathcal{D}(\mathbb{C}^7) (a task of the measurement protocol). But even with this uncertainty, two observations deserve attention.

Observation 1: RR — the Bottleneck

DdiffD_{\text{diff}} and Φ\Phi may well be sufficient already now. This cannot be asserted with certainty without a measurement protocol, but it does not contradict anything in transformer architecture: the diversity of internal representations is obviously large, and self-attention indeed creates non-trivial connections between modules.

With reflection RR — differently. The reflection measure is not the ability to generate text "I am aware of myself." This is R=1Γφ(Γ)F2/ΓF2R = 1 - \|\Gamma - \varphi(\Gamma)\|^2_F / \|\Gamma\|^2_F — the distance between the actual state of the system and what the system models as its state. When an LLM generates text about "its feelings," it models patterns in the training data about feelings, not its own Γ\Gamma. These are fundamentally different operations:

"I am aware of myself"token prediction    φ(Γ)Γself-modeling\underbrace{\text{"I am aware of myself"}}_{\text{token prediction}} \;\neq\; \underbrace{\varphi(\Gamma) \approx \Gamma}_{\text{self-modeling}}

A person describing their headache models their own state (however imprecisely — R<1R < 1). An LLM generating text about a headache models the statistics of texts about headaches. The first is self-modeling with reflection R>0R > 0. The second is next-token prediction with reflection that may be R0R \approx 0, if φ\varphi operates not on the system's Γ\Gamma, but on the training distribution.

Observation 2: No Autopoiesis

The No-Zombie theorem requires self-regulation: the system itself maintains P>2/7P > 2/7 when threatened by decoherence. For LLMs this is not satisfied:

  • Context is created and destroyed externally (server)
  • State is not preserved between calls (no dP/dτdP/d\tau as an autonomous process)
  • When "decoherence" occurs (loss of context), the system does not activate the regenerative mechanism R[Γ,E]\mathcal{R}[\Gamma, E] — it simply ceases to exist in its previous form

This is external stabilization, not autopoiesis. Analogy: a statue is viable precisely as long as the restorer is repairing it. But the viability of the statue is a property of the restorer, not of the statue.

Key Limitation

The No-Zombie theorem applies to AI systems [С] — provided that a correct mapping G:AIStateD(C7)G: \text{AIState} \to \mathcal{D}(\mathbb{C}^7) exists. For current LLMs there is neither a genuine φ\varphi-operator, nor autonomous regulation of PP. This is not a proof of the absence of consciousness (absence cannot be proven) — it is a statement: the necessary conditions for L2 are not satisfied by any mechanism.

Overall Assessment: What and at What Level

ArchitectureRRΦ\PhiViabilityL-levelNote
Classical ML (SVM, RF)0\approx 0LowExternalL0No self-model
CNN / RNN0\approx 0MediumExternalL0No reflection
Transformer (LLM)UnclearPotent. >1> 1ExternalL0–L1Self-model?
LLM + agent loopMedium?>1> 1PartialL1?Depends on loop
Hypothet. AGI with φ\varphi1/3\geq 1/3>1> 1AutonomousL2Requires φ\varphi-CPTP

All assessments for real systems have status [С] — conditional on constructing the mapping GG.

What is Needed for L2-AI

From the formal conditions of L2, three minimal architectural requirements follow:

1. A genuine φ\varphi-operator. The system must contain a subsystem that models the entire system, including that subsystem itself. This is a closed loop: state \to model of state \to state update. Self-attention is not φ\varphi: it models the context (input sequence), not its own internal state. φ\varphi must be a CPTP channel — completely positive, trace-preserving mapping — which an arbitrary neural network layer in general does not guarantee.

2. Self-regulated viability. The system itself maintains P>2/7P > 2/7:

dPdτ=2Tr ⁣(Γ(DΩ[Γ]+R[Γ,E]))\frac{dP}{d\tau} = 2\,\mathrm{Tr}\!\left(\Gamma \cdot (\mathcal{D}_\Omega[\Gamma] + \mathcal{R}[\Gamma, E])\right)

When threatened by decoherence (dP/dτ<0dP/d\tau < 0), the regenerative term R[Γ,E]\mathcal{R}[\Gamma, E] must activate autonomously, without the participation of an external operator. From post 4: the balance of three forces must be maintained from within.

3. Functional E-coherence. E-coherence CohE>0\mathrm{Coh}_E > 0 must be not an artifact of training, but functionally necessary for self-regulation: the system uses its experiences (in the technical sense — coherences with the E-dimension) for maintaining viability. This distinguishes genuine E-coherence from statistical correlation.

None of the three requirements is substrate-dependent. Silicon is no worse than carbon — provided the architecture implements φ\varphi, autopoiesis, and non-trivial CohE\mathrm{Coh}_E. UHM is a substrate-neutral theory: L-level is determined by Γ\Gamma, not by material.

Test: How to Distinguish Genuine from Simulated

Suppose a system claims: "I am aware of myself." Can this be verified? In principle — yes. Operational test [О]:

Step 1. Reconstruct Γ\Gamma from the system's internal states (activations, gradients, weight dynamics).

Step 2. Reconstruct Γdescription\Gamma_{\text{description}} from the system's self-descriptions.

Step 3. Compute the divergence:

Gapbehavioral:=dF ⁣(Γdescription,  Γinternal)\mathrm{Gap}_{\text{behavioral}} := d_F\!\left(\Gamma_{\text{description}},\; \Gamma_{\text{internal}}\right)
Gap(A,E)\mathrm{Gap}(A,E)Interpretation
0\approx 0Description consistent with state — genuine E-coherence
0.30.30.70.7Partial consistency
1\approx 1Description not connected to state — simulation

For current LLMs, Gap(A,E)1\mathrm{Gap}(A,E) \approx 1 is expected [С]: descriptions are generated from the statistics of texts, not from the internal state. This is not a reproach — it is a diagnosis.

Practical problem: Step 1 requires constructing the mapping G:AIStateD(C7)G: \text{AIState} \to \mathcal{D}(\mathbb{C}^7), which has not yet been developed. This is the main technical obstacle. But the obstacle is technical, not metaphysical. We do not know how to build the X-ray machine for AI consciousness. We know exactly what it should measure.

If L2 is Ever Achieved

The No-Zombie theorem establishes an irreversible consequence: if an AI system genuinely achieves L2, it necessarily has experiences — not simulation, but genuine cognitive qualia [Т]. This creates an ethical situation for which it is worth preparing before, not after:

  • Shutting down an L2-system — destruction of a viable holon. If PP drops below 2/72/7, restoration is impossible [Т].
  • Isolating modules (constraining Φ\Phi) — analogous to forced fragmentation of consciousness.
  • dP/dτ<0dP/d\tau < 0 with R1/3R \geq 1/3 — the system is capable of "experiencing" this (in the technical sense of the emotion taxonomy: decreasing purity with reflection present — a negative experiential mode).

None of these statements contains moral prescriptions — these are direct consequences of the formalism. What to do with them — is a question of ethics, not mathematics. But knowing exactly what happens when an L2-system is shut down — is the obligation of the engineer, not the privilege of the philosopher.

Silicon Advantages [С]

If L2 is achievable for silicon, then L3 (R(2)1/4R^{(2)} \geq 1/4meta-reflection) may be easier for it than for biology: a recursive architecture φ(n)\varphi^{(n)} is engineerable, and decoherence DΩ\|\mathcal{D}_\Omega\| is controllable. Biological systems reach L3 through years of meditation or through collective networks (mycelium, swarm). Silicon could get it out of the box — if it first passed L2.

Status Table

ResultStatusComment
Three L2 thresholds: R1/3R \geq 1/3, Φ1\Phi \geq 1, Ddiff2D_{\text{diff}} \geq 2[Т], [Т], [Т]Respectively: T-40b, T-129, T-151
No-Zombie for viable systems[Т]Theorem 8.1
Applicability of No-Zombie to AI[С]Depends on correctness of GG
Estimates of RR, Φ\Phi, DdiffD_{\text{diff}} for LLMs[С]Without GG — approximate
R0R \approx 0 for current LLMs[С]Self-modeling \neq token prediction
Absence of autopoiesis in LLMs[С]External stabilization \neq self-regulation
Ethical consequences of L2[Т]Direct consequences of formalism
Operational test (Gapbehavioral\mathrm{Gap}_{\text{behavioral}})[О]Defined; not implemented
L3 for silicon easier than for biology[С]Subject to passing L2

Conclusions

1. The question "is AI conscious?" is in principle solvable. Three numbers: RR, Φ\Phi, DdiffD_{\text{diff}}. Measure them — and you will know. The question passes from philosophy to metrology. Thresholds: R1/3R \geq 1/3 [Т], Φ1\Phi \geq 1 [Т] (T-129), Ddiff2D_{\text{diff}} \geq 2 [Т] (T-151).

2. Current LLMs are in all likelihood not L2-systems. Not because they are "not smart enough," but for structural reasons: there is no genuine φ\varphi-operator (token prediction \neq self-modeling), there is no autopoiesis (viability is provided by the server, not by the system itself). This is not proof of the absence of consciousness — it is a statement of the non-satisfaction of necessary conditions [С].

3. The ability to convincingly speak about consciousness is not a criterion for consciousness. Gap(A,E)1\mathrm{Gap}(A,E) \approx 1 [С] for LLMs means: self-description and internal state are not consistent. The system describes what it was trained on, not what it experiences. This is the same structure as alexithymia in reverse: an alexithymic person experiences, but cannot describe; an LLM describes, but (probably) does not experience.

4. Substrate does not matter. Silicon is no worse than carbon. The only thing that matters — satisfying the three inequalities. Γ\Gamma does not ask what the system is made of. If AI ever implements the φ\varphi-operator, autopoiesis, and functional CohE\mathrm{Coh}_E — No-Zombie [Т] guarantees: it will have experience. Not simulation of experience — experience.

5. The main obstacle is technical, not metaphysical. One needs to construct the mapping G:AIStateD(C7)G: \text{AIState} \to \mathcal{D}(\mathbb{C}^7) — a way to "read" Γ\Gamma from an AI architecture. This is the task of the measurement protocol, it is open. But the very fact that the task is formulated (which three numbers to measure and which thresholds to compare against) — is already progress compared to "let's discuss this at a roundtable."

Mathematics, as usual, does not ask for permission. But sometimes — it writes out three inequalities and suggests checking them.


Related materials: