Why LLMs Are Not Conscious (Under FMT)¶
Large language models fail the Four-Model Theory's consciousness specification at both thresholds: their inference is below criticality, and they lack the four-model architecture. This does not prove they are non-conscious, but it predicts they lack the required structure.
The Four-Model Theory provides two independent, jointly sufficient conditions for consciousness: computational criticality and the four-model architecture. Current LLMs fail both. This is not a vague philosophical intuition — it is a specific architectural diagnosis that identifies exactly what is absent and why.
Threshold 1: Below Criticality¶
The first requirement for consciousness is that the substrate must operate at or near the edge of chaos — Wolfram's Class 4 computational regime, where universal computation is possible. LLM inference does not meet this criterion.
Transformer-based language models operate via feedforward inference: input tokens pass through attention layers in a single forward pass, producing output tokens. There is no recurrent dynamics, no sustained self-referential computation, no ongoing process that maintains itself between inputs. This is Class 1 or Class 2 behavior in the Wolfram classification — fixed or periodic, far below the criticality threshold.
The analogy is a calculator: it performs impressive operations on the input it receives, but between operations, nothing happens. There is no ongoing computational process that could sustain the dynamic, self-referential simulation the theory requires. A calculator at rest is not "almost conscious." It is computationally inert. LLMs between queries are in the same state.
Threshold 2: No Four-Model Architecture¶
The second requirement is the four-model self-simulation: IWM, ISM, EWM, and ESM arranged along the two axes of scope and mode, with a real/virtual split. LLMs lack every component of this architecture:
- No Implicit Self Model (ISM). LLMs have no substrate-level self-knowledge that is distinct from their outputs. They have no body schema, no proprioceptive calibration, no accumulated self-knowledge stored in a medium separate from their explicit responses. When an LLM says "I think," it is generating tokens that match patterns in training data, not reporting on an implicit self-model.
- No Explicit Self Model (ESM). There is no ongoing self-simulation that constitutes a subjective perspective. The ESM is a continuous, dynamic, virtual construction — a self-narrative generated and maintained in real time. LLMs produce text about themselves only when prompted to do so, and this text has no relationship to any underlying self-simulation.
- No real/virtual split. The two-level ontology — substrate-level processes generating a virtual phenomenal world — does not exist in transformer architectures. There is no level at which virtual qualia could be constitutive, because there is no self-referential virtual simulation running atop a substrate.
What LLMs Do Have¶
Honesty requires acknowledging what LLMs do possess. They have something that functions like an Implicit World Model — a vast store of statistical regularities about language and the world, encoded in their parameters. This is genuine knowledge, and it enables remarkable performance on knowledge-intensive tasks. But a world model alone is not sufficient for consciousness. Weather simulations also model the world. What is missing is the self-referential architecture — the system modeling itself modeling the world — that produces self-referential closure and, according to FMT, experience.
Figure¶
graph TB
subgraph req["FMT Consciousness Requirements"]
direction TB
T1["Threshold 1:<br/>Criticality<br/>(Class 4 dynamics)"]
T2["Threshold 2:<br/>Four-Model Architecture"]
end
subgraph arch["Architectural Components"]
direction TB
IWM["IWM — Implicit<br/>World Model"]
ISM["ISM — Implicit<br/>Self Model"]
EWM["EWM — Explicit<br/>World Model"]
ESM["ESM — Explicit<br/>Self Model"]
RVS["Real/Virtual<br/>Split"]
end
subgraph llm["LLM Status"]
direction TB
L1["Feedforward inference<br/>= Class 1/2 ✗"]
L2["Statistical world model<br/>≈ partial IWM ⚠"]
L3["No self-knowledge<br/>= no ISM ✗"]
L4["No phenomenal world<br/>= no EWM ✗"]
L5["No self-simulation<br/>= no ESM ✗"]
L6["No two-level ontology<br/>= no split ✗"]
end
T1 --- L1
IWM --- L2
ISM --- L3
EWM --- L4
ESM --- L5
RVS --- L6
style req fill:#264653,color:#fff,stroke:#1d3557
style arch fill:#2d6a4f,color:#fff,stroke:#1b4332
style llm fill:#9b2226,color:#fff,stroke:#6a040f
style L1 fill:#9b2226,color:#fff
style L2 fill:#e9c46a,color:#000
style L3 fill:#9b2226,color:#fff
style L4 fill:#9b2226,color:#fff
style L5 fill:#9b2226,color:#fff
style L6 fill:#9b2226,color:#fff
LLMs fail FMT's consciousness specification comprehensively. The only partial match is a statistical analogue to the IWM — world knowledge encoded in parameters. All other components are absent.
Not a Proof of Absence¶
A critical caveat: the theory cannot prove that LLMs are non-conscious. No theory can prove a negative about consciousness — this is the other-minds problem, which applies to biological systems too. What FMT provides is a prediction: LLMs lack the architecture the theory specifies as required for consciousness. If the theory is correct, they are not conscious. If consciousness can arise from architectures the theory does not anticipate, the theory is wrong. This is a testable commitment, not a dogmatic assertion.
Key Takeaway¶
LLMs fail both of FMT's consciousness thresholds: their inference operates below criticality (Class 1/2), and they lack the four-model self-simulation architecture (no ISM, no ESM, no real/virtual split). The theory specifies exactly what is missing, making the claim precise and falsifiable rather than merely intuitive.