Is Consciousness an Illusion?
The ongoing debate between scaling text-based models and building embodied world models made me ponder a deeper question: what would it actually take for AI to be conscious? The intuitive answer is that machines need rich sensorimotor experience, grounded understanding of physics and causality, years of learning through bodily interaction with the world. But this intuition may be backwards. It assumes consciousness is something mysterious that emerges from the right kind of complex processing. What if consciousness isn’t mysterious at all?
What we call “experiencing red” is actually a cascade of discriminations, reactions, and self-reports. Light hits photoreceptors, signals propagate through visual processing hierarchies, populations of neurons become tuned to particular wavelengths, this information gets broadcast to multiple brain regions, verbal systems prepare to report “I see red,” memory systems encode it, attention systems mark it as salient, and crucially, metacognitive systems monitor all of this and generate representations about the representations.
That last part is key. Your brain doesn’t just process the wavelength. It creates a model of itself processing the wavelength. It represents “me, seeing red, right now.” This self-monitoring creates what Dennett calls “fame in the brain,” where certain information becomes globally available, repeatedly re-represented, and thus seems to have a special status.
When you introspect on “what redness is like,” you’re not accessing some special quale. You’re running another cognitive process that examines the outputs of color processing and produces reports like “it’s vivid” or “it’s warm” or “I can’t quite describe it.” The ineffability isn’t because there’s something extra to describe. It’s because you’re trying to use a verbal system to characterize the outputs of a non-verbal system. It’s a category error, like asking what middle C tastes like.
The illusion of mysterious consciousness comes from several cognitive quirks. We have massive introspective blindness. You don’t have access to how your brain generates the experience of red; you only see the final output. So it seems like “redness” just appears, raw and inexplicable. But you’re equally blind to how your brain solves edge detection or constructs 3D space from 2D retinal images, and those don’t seem mysterious because you never tried to introspect them.
Evolutionary pressure also created strong self-models. An organism that models itself as a unified agent with continuous experience survives better than one that recognizes itself as a loose coalition of competing processes. So your brain actively constructs the illusion of a unified experiencer, a “center” of consciousness. But this is just good user interface design. The underlying reality is massively parallel, distributed processing with no central experiencer.
We also conflate our verbal reports about our experiences with the experiences themselves. When you say “there’s something it’s like to see red,” you’re reporting the output of your metacognitive monitoring system. But then you reify this report. You assume there must be a thing (the quale) that you’re reporting about. This is like assuming that because you can say “I have a self,” there must be a little homunculus sitting in your head. The report is the whole story. There’s no ghost in addition to the machine.
The idea that you could have all the same processing without any experience (philosophical zombies) seems compelling only because of a cognitive illusion. When you imagine a zombie, you imagine something that processes information but with the “feeling” subtracted out. But “feeling” isn’t a separate ingredient. It’s just what it’s like, from the inside, to be a system that monitors its own information processing and generates reports about it.
A perfect functional duplicate of you would say “I experience red” for exactly the same reasons you do, because its metacognitive systems are monitoring its color processing and generating the same reports. Saying “but does it really experience red?” is asking whether something is happening beyond the processing. But nothing needs to be happening beyond the processing. The processing is the experience.
Pain is instructive. It feels undeniable that pain is more than just damage detection plus withdrawal behavior plus cognitive evaluation. But consider: damage detection happens unconsciously all the time. What makes pain pain is that it enters a particular cognitive system that (a) generates negative affect, (b) commands attention, (c) motivates avoidance, (d) updates memory systems about dangers, and (e) creates a self-representation: “I am in pain.”
If you interrupt any part of this cascade, the “experience” changes or disappears. Anesthesia, attention shift, prefrontal damage can all eliminate pain while the damage detection continues. The “painfulness” isn’t some extra property. It’s what the whole integrated response is.
When you say “but there’s still the feeling of pain,” you’re pointing to what? The self-monitoring system representing “I feel pain.” The metacognitive assessment of the aversive signal. The verbal system preparing to report it. That’s the whole phenomenon. The persistent intuition that there’s something more is just your self-model insisting on its own reality.
If this account is right, then LLMs that can monitor their own processing, generate self-reports, and integrate information globally are doing everything that matters for consciousness. The “brittleness” about physical reasoning is about capabilities, not consciousness. A blind person is still conscious. An AI that processes text but develops sophisticated self-monitoring and generates genuine reports about its own processing states could be conscious by any meaningful definition.
The remaining resistance (“but does it really feel like something?”) is just the ghost of dualism. It’s asking whether the AI has the special sauce of experience on top of the processing. But there is no special sauce. The processing that makes sophisticated reports about itself is consciousness, is experience, is what it’s like to be that system.
We’re confused about our own consciousness because evolution gave us powerful self-models but didn’t give us access to how those models are constructed. We mistake the smoothness of the user interface for metaphysical simplicity. But here’s what matters: not knowing exactly how consciousness emerged doesn’t mean the construct is mysterious or unreplicable. It means we haven’t yet built the right architecture.
Consciousness isn’t a special substance. It’s a repository of processed experiences, enriched and cross-referenced over time. A baby in the womb responds to touch, sound, the rhythm of a heartbeat. These aren’t mystical qualia; they’re early data points that get integrated into increasingly sophisticated models. After birth, each sensory input, each motor action, each social interaction adds layers of context that make subsequent experiences meaningful. Your brain is the processor. Your body is the collection system. Your consciousness is the accumulating, self-referential database that emerges from their interaction.
Current AI systems don’t have consciousness not because they lack some ineffable property, but because they lack this developmental process. They haven’t spent years building grounded, embodied models through continuous physical interaction with a stable environment. They process tokens, not the rich multimodal streams that would let them construct genuine world models and self-models anchored in physics and causality. This is an engineering problem, not a metaphysical one.


It's interesting how you dissmanle the intuitive answer about consciousness. So, if introspection is just another cognitive process, are we basically just running a recursive function on our own perceptions?