Why Chalmers' thought experiments dissolve under scrutiny, and what a self-referential account of consciousness actually predicts
A word, stripped to its basics, is a sound with a use. "Water" is a noise that functions in contexts: requesting it, describing it, using it as a verb. The meaning is not hidden behind the word waiting to be uncovered. It is constituted by the uses themselves, the requests, the descriptions, the comparisons, the daily interactions that built up the word's role over time.
This seems obvious until you notice how much philosophy ignores it.
When someone asks "but what is water, really?" — not as a request for the chemical formula, not asking how to use the word more precisely, but asking as if the word contains some deeper truth about essence — something has already gone wrong. The question strips the word from every context that gave it meaning and then treats the resulting emptiness as a profound mystery. It mistakes a contextless sound for a window onto some objective, context-independent category.
Wittgenstein called this a language going on holiday. The word leaves its working context and starts being asked to do something it was never built for.
The test is simple.
01 -- Look at how a word actually functions. What contexts does it appear in? What work does it do? What questions does it answer?
02 -- Ask whether the philosophical question has removed that context entirely. Has the word been treated as if it names some Platonic object with an essence independent of its use? Has an ongoing process been frozen into a static thing that can then be interrogated about its nature?
"Is water wet?" sounds coherent. But wet and dry are properties things have in relation to water. Nobody ever defined what wetness applied to water itself means. The question has grammatical structure and no content.
"What is the essence of life?" Same move. Life is a category we apply to a collection of behaviours, processes, and relations. Ask what life is stripped of all those and you have not posed a deep question. You have removed the thing you were asking about.
"What is freedom, really?" Freedom only has meaning in relation to constraints. Strip the constraints and the word has nothing to push against.
This is not scepticism about whether these things exist. Water exists. Life exists. Freedom exists. Consciousness exists. The point is that they exist as relational, processual, context-dependent phenomena — not as Platonic objects with intrinsic essences waiting to be discovered by sufficiently careful introspection or philosophical analysis.
Every thought experiment in the philosophy of consciousness that has generated decades of apparently intractable debate makes exactly this move. It takes a relational, processual phenomenon and treats it as a discrete object that can be present or absent, swapped out, or imagined away while everything functional stays fixed.
The debates feel intractable not because the questions are deep but because the questions are malformed. A malformed question cannot be answered. It can only be dissolved.
Before any neuroscience, before any formal argument, experience itself has something to say.
Try this right now. Notice what is happening as you read this sentence. The words arrive. Meaning assembles without your instruction. You did not choose what to understand next — understanding just happened. Your eyes moved down the page without a decision to move them. Somewhere in the background, a thought about something unrelated surfaced briefly, or your posture shifted, or you became briefly aware of a sound you had been ignoring. None of that was chosen. It arrived.
This is not an edge case. This is the texture of experience at every moment.
Descartes, sitting at his desk with a candle beside him, ink pot to hand, the faint smell of the room around him, was doing exactly this as he wrote his famous arguments. Every concept he reached for — substance, thought, existence, doubt — arrived pre-assembled. The word "justice" does not arrive as a careful enumeration of its components. It arrives as a dense compressed node of associations built up over decades: vague feelings, remembered arguments, contexts in which the word was used, a relational complexity so high it cannot be consciously unpacked, retrieved in an instant precisely because it has been thought about so many times that the mind has made it automatic. He did not think through all of that. He retrieved it.
He attended to one syllogism at a time. The rest of his experience ran as background: the candle, the cold, the feel of the quill, the structure of the room assembled from memory rather than carefully recomputed at each moment. His confidence in each logical step was itself a feeling — a sense of rightness, not a bottom-up calculation. The certainty he felt was an emotional state accompanying the reasoning, not a separate verification process running alongside it.
And his attention was unbidden. Even as he intended to think carefully and systematically, each thought that surfaced into the foreground of his awareness arrived without prior announcement. The intention itself was unbidden. Where did the decision to decide come from?
All of this — the automated retrieval, the fragmented attention, the constructed confidence, the unbidden thoughts, the background processing, the room assembled from model rather than perception — was happening in a man arguing for a unified, persistent, non-physical soul that transparently knows its own contents.
The argument was falsified by the process generating it.
Hume put it plainly:
"For my part, when I enter most intimately into what I call myself, I always stumble on some particular perception or other, of heat or cold, light or shade, love or hatred, pain or pleasure. I never can catch myself at any time without a perception, and never can observe any thing but the perception."
Not a self. Just perceptions in succession.
The now is thin. Continuity is not experienced directly — it is assembled by memory after the fact. Remove the memory and it collapses.
What experience actually shows, when examined rather than assumed, is construction all the way down. Fragments, automated processes, inferred continuity, retrieved concepts arriving pre-compressed, attention that moves without instruction, confidence that is feeling rather than calculation.
The hard problem assumes experience is a vivid, unified, continuous thing whose felt quality demands special explanation. That assumption is already wrong. The thing requiring explanation is considerably thinner than advertised. Much of what felt like mystery dissolves once the inflation is corrected.
This exercise demonstrates constructedness. It does not by itself dissolve the Chalmers intuition — the vantage point argument handles that separately.
Start with the eye.
Light hits the retina and the processing that follows is immediately relational. Photoreceptors transduce wavelengths, but opponent-process cells encode colour as contrast — red defined by the suppression of green, blue by the suppression of yellow. What leaves the eye is not raw wavelength data. It is already a set of relations.
This continues upward. The lateral geniculate nucleus gates and streams the signal. V1 extracts edges and orientations. V2 separates figure from ground. V4 computes colour constancy — red looks red under different lighting conditions because V4 normalises against surrounding colours, not against the wavelength itself. By the time the signal reaches the temporal cortex for object recognition, what is being processed is not a physical measurement. It is a relational signature, abstracted across multiple transformations, none of which are accessible to awareness.
In parallel: the dorsal stream connects visual processing to motor planning and spatial cognition. The superior colliculus integrates visual, auditory, and somatosensory maps — a sound to the right shifts visual attention automatically, beneath awareness. The amygdala assigns threat-relevant valence via a fast subcortical route, before full object recognition completes. The insula integrates incoming signals with body state — heart rate, muscle tension, gut — continuously, beneath awareness.
None of this surfaces. It is all in the dark.
By the time anything reaches the prefrontal cortex and the network of areas involved in self-referential processing, it has already been: colour-normalised, edge-extracted, figure-ground separated, object-recognised, spatially located, motion-processed, threat-evaluated, valence-assigned, motor-prepared, and cross-referenced against body state, prior encounters, and emotional history.
The self-referential loop encounters the heavily compressed output of all of that simultaneously. It never sees a wavelength. It never sees an edge. It sees red-apple-familiar-safe-reachable-slightly-hungry, all at once, as a single relational gestalt.
Now consider what the self-referential loop actually does. It does not passively receive this gestalt. It compares it against its own prior states: what working memory was holding moments ago, what long-term memory reconstructs of prior encounters, what the body's current condition is, what the predictive machinery expected to be present. The loop cross-compares the current gestalt against every other accumulated relation simultaneously. Nothing arrives neutrally. Everything lands in a context the system itself built.
This is not feedforward processing. A feedforward network transforms inputs and passes them on. Each layer preprocesses for the next. Nothing refers back. Nothing persists. There is no vantage point because there is nothing accumulating a perspective.
The self-referential loop is different in kind, not just degree. Current processing is continuously compared against the system's own prior states. Interoceptive signals feed into the same recurrent architecture, so body state and sensory input are integrated rather than parallel. Working memory holds representations across seconds. Long-term memory reconstructs prior encounters, shaped by current state. Dopaminergic systems have sculpted, over a lifetime, which relations got strengthened and which faded. The system's current processing is never just processing incoming signals. It is always processing relative to everything it has accumulated.
Top-down, the picture runs in reverse. Conceptualisations, language, emotion concepts, self-narratives recursively shape what gets processed and how. The concept of red is not a neutral label applied after perception. It is a dense node of associations — objects, memories, valences, contexts — that actively shapes the processing of chromatic input. Detrain those conceptual layers through sustained meditation or deliberate behavioural change and they dissolve back into simpler relational processing beneath them. The layers were constructed, not discovered.
The loop has no access to any of the mechanisms described above. It encounters only their outputs. It can only compare representations against other representations. It hits a floor — not because something mysterious is there, but because the floor is the structural limit of a system that can only refer to its own outputs.
That floor is what philosophers call ineffability. And the experience of redness — rich, immediate, irreducible — is what it is to be that loop encountering that gestalt, in relation to everything else the system has ever accumulated. Those are not two separate facts requiring separate explanation. They are the same fact, from inside and outside respectively.
A convergent body of neuroscientific research points toward partial overlap in the mechanisms emphasized by different theories. Five major theories of consciousness—recurrent processing theory, integrated information theory, global neuronal workspace theory, predictive processing and neurorepresentationalism, and dendritic integration theory—show convergence around recurring mechanistic motifs, despite disagreement on many details; these include the widespread involvement of recurrent processing, large-scale integration across brain systems, and the association of conscious states with the preservation of these dynamics, and their disruption across anesthesia, dreamless sleep, and disorders of consciousness.
Recurrent processing theory shows directly that feedforward sweeps reaching early visual cortex within 40 milliseconds are insufficient for conscious experience. It is only when sustained recurrent interaction between early and higher areas emerges after around 100 milliseconds that experience arises. Interrupt that recurrence with transcranial magnetic stimulation and the stimulus disappears from awareness despite the feedforward pass completing normally.
Dendritic integration theory adds cellular specificity. Layer 5 pyramidal neurons serve as a nexus point for both cortico-cortical and thalamocortical information flow, with apical and basal dendritic compartments integrating top-down context and bottom-up input simultaneously. General anesthesia decouples these compartments, collapsing the thalamocortical loops that sustain conscious processing. Consciousness tracks the coupling state of these neurons across wakefulness, sleep, and anesthesia with remarkable precision.
Predictive processing and neurorepresentationalism frame conscious experience as a multimodal, hierarchically integrated world model — not a passive readout of sensory input but an active inference process comparing predictions against incoming signals across multiple levels simultaneously. This maps directly onto the self-referential loop's cross-comparison of current states against accumulated representations.
These theories disagree on what exactly is sufficient for consciousness. Recurrent processing alone appears necessary but not sufficient — recurrence is ubiquitous in neural systems that are not conscious. Not every recurrent loop is enough. The architecture matters.
The hard problem is typically defended through thought experiments: the philosophical zombie, Mary's room, inverted qualia. Each is presented as a demonstration that phenomenal experience is something over and above any physical or functional description. Each makes the same error identified in section one: treating a relational, processual phenomenon as a discrete object that can be present or absent, swapped out, or imagined away while everything relational stays fixed.
When you describe another person's experience, you attribute it to their system — and that attribution is epistemically all you have. Their experience does not surface to you. You assume it exists by analogy with your own, but you cannot access it from outside their vantage point. This is not a limitation of current science. It is structural. Any claim about another system's experience necessarily describes it as a process relative to that system, not as an object you can inspect, subtract, or stipulate away.
The moment you describe Tabitha's experiences, you have already located them processually at Tabitha. There is nothing to subtract.
A philosophical zombie is stipulated to be physically and functionally identical to a conscious human but without subjective experience. The claim is that this is conceivable, and conceivability implies possibility, which implies consciousness is something over and above the physical-functional description.
Try to construct this scenario carefully.
You imagine Tabitha. You imagine her architecture in full. Now you try to imagine the same Tabitha without experience. What happens? One of two things.
You occupy her vantage point — you imagine being Tabitha, seeing through her eyes. In this case you have not demonstrated separability. You have run her perspective. The lights are on because you turned them on by stepping inside.
Or you step outside and take a third-person stance. You describe her architecture from outside and stop attributing experience to it. The lights appear to go off.
But the lights going off is not a fact about Tabitha's architecture. It is a fact about your descriptive stance. You switched from attributing experience to not attributing it. Two third-person descriptions of the same architecture — one that attributes experience and one that does not — are not descriptions of two different systems. They are the same description with a different attribution. The zombie thought experiment mistakes a shift in descriptive stance for a coherent alternative state of the system.
You cannot imagine not being conscious either. Any attempt to do so is itself more conscious processing. You can take a third-person stance on your own brain — imagining it in the dark, neurons firing without anyone home — but you have already switched to a third-person description of yourself. You have not conceived of a first-person absence. You have conceived of a third-person description that omits attribution.
This is not an epistemic limitation. The concept itself is empty. Conceivability requires coherent content, and the zombie has none.
Mary knows every physical fact about colour vision but has never seen red. When she finally does, does she learn something new?
Mary does not gain a new fact. She gains a new representational instantiation. Her self-referential loop now has red as direct input rather than as described input. What changes is not her knowledge base but the way the information is processed relative to everything else the system has accumulated. The vantage point shifts.
Jackson's argument requires this shift to count as a knowledge gap in the physical account. But expecting Mary's new representational instantiation to count as missing knowledge is expecting first-person access to be derivable from third-person description. That expectation is a restatement of what a vantage point structurally is. It is not an argument against physicalism.
Chalmers' fundamentality move does not retain pain as an explanandum — it evacuates everything constitutive of what pain is and keeps the label. At that point he and the mechanistic account are no longer disagreeing about pain. One talks about stable intrinsic properties of the universe, and the other talks about subjective processes, with only the word being shared.
Two systems could be functionally identical while one experiences red where the other experiences green. Since the inversion is undetectable, phenomenal character must float free of functional organisation.
Red and green are not independent qualities assigned to wavelengths. They are opponent channels — neurally defined by mutual suppression. Two systems with identical relational architecture cannot differ phenomenally. There is nothing left over for the inversion to operate on. The thought experiment assumes phenomenal character is separable from relational structure. That is the contested premise, not an established starting point.
There is also a double standard worth naming. The demand for logical closure from a mechanistic account of consciousness is not a principled philosophical requirement — it is selectively applied. At what point would any third-person description be sufficient? The demand can always be pushed back one step regardless of how complete the account is. That is not a feature of consciousness specifically. It is a feature of the vantage point difference between first and third person descriptions — which the framework already accounts for. Either consciousness is explained via an inner homunculus — which multiplies entities without explanatory gain — or the self-referential process within the brain simply constitutes consciousness.
Every thought experiment follows the same structure. Stipulate a system. Then stipulate that the phenomenal character is different or absent while everything relational stays fixed. Conclude that phenomenal character must be something additional. But the middle stipulation is never examined. It assumes separability from the outset — the very thing the argument is supposed to establish. Each thought experiment is circular.
Dissolution is not a consolation prize. Every alternative self-destructs on contact with the category error. The mechanistic account inherits the ground by default as the only position that does not undermine itself.
The objection that this account is equally question-begging — that it assumes non-separability rather than establishing it — gets the dialectic backwards. Separability requires treating phenomenal character as a discrete object that could be present or absent independently of relational structure. That is the category error identified in section one. The burden is on the separability advocate to specify what phenomenal character is, independently of any relational description, without using vantage-point relative language. That specification has never been given.
A reader might resist the full claim — that the self-referential loop just is the vantage point. Fine. Grant nothing beyond what the mechanistic account literally describes.
You would still have to grant: the self-referential loop has a quality of irreducibility in a specific, non-mysterious sense. It can access only the outputs of the layers beneath it. It hits a floor. That floor is structural, not mysterious.
And you would have to grant: the loop can only relate every representation against every other representation. The chromatic state that constitutes red is defined in relation to every other accumulated state simultaneously. Tracked frame by frame with state persistence, this is a system where nothing arrives as an isolated signal. Everything lands in relation to everything else the system has built.
A feedforward network cannot do this. Not because it lacks complexity but because it lacks the architecture. No feedforward system has a current state relative to its own prior states.
Whatever you want to call what the self-referential loop has, it is already something categorically different from feedforward processing. The ineffability and the experience — the floor and the richness — are both direct consequences of that architecture. They are not two separate features. They are the same architectural fact described from two directions.
A system with sufficiently rich relational architecture — genuine recurrent self-reference, persistent interoceptive integration, accumulated cross-modal representations with a structural floor — would report exactly what any conscious system reports: irreducible abstractions it cannot get below, fleeting states only describable relative to other states, not because it was trained to produce those outputs but because that is what the architecture generates when turned toward itself. Current large language models, which predict tokens without any of this architecture, are not that system.
If the argument above is correct, a question remains. Why does the hard problem feel so compelling? Why have careful, serious philosophers defended it for decades without feeling the force of the dissolution?
Thomas Metzinger's self-model theory of subjectivity offers a partial answer, though he does not draw the full implication.
Metzinger argues that the brain generates a phenomenal self-model — a representation of the system by the system. Crucially, this model is transparent: you do not see it as a model. You see through it. The self is not experienced as a representation of a self. It is experienced as direct reality.
In a related argument, Metzinger proposes what he calls necessary self-reification: the brain stops recursive self-modelling by presenting the ongoing self-referential activity as an object. The process gets frozen into a thing. The self-model becomes, phenomenologically, a self.
The implication he does not fully draw is this: the hard problem is an artefact of this architectural feature, linguistically confused. Specifically, the experiential (subjective; first-person) and descriptive (objective; third-person) 'language games' are products of two separate vantage points, but the experiential gestalt gets reified as an object, and hence 'imagined' as a separable property. The conflation between the two lanugage games results in a linguistic confusion as specified in Section I.
When you introspect, you do not encounter an ongoing self-referential process. You encounter something that presents as a discrete, unified, persisting thing with intrinsic properties. Metzinger he architecture delivers experience as object-like because that is how the self-model is presented. Philosophical intuitions about phenomenal properties being separable, intrinsic, and mysteriously resistant to physical explanation are not arbitrary. They are downstream of this presentation.
Chalmers is not making a simple error. He is taking seriously the deliverances of a system that presents its own processes as objects. The hard problem feels ineliminable because the confusion is generated at the architectural level, before philosophical reflection begins. Introspection cannot correct it because introspection operates through the same self-model that produces it.
The intuitions are real. What they track is the brain's own subject-forming presentation of its processes, not a genuine gap in the physical account.
Dissolving the hard problem is not dissolving all questions about consciousness. Three survive cleanly.
First: why did self-referential architecture evolve at all? The selective pressure is clear — a system that models its own states and integrates that against body and environment is vastly more flexible than one that does not. But the precise mechanistic history, which intermediate forms enabled the transition and across what timescales, remains empirically open. This is a question for evolutionary biology, not philosophy.
Second: how precisely do the mechanisms work? The recurrent processing and thalamocortical evidence points in the right direction, but the exact dynamics by which gestalts surface, how cross-modal integration binds into a unified moment, and how the self-referential loop's specific architecture produces the structural floor — these remain open empirical questions. The philosophical framework predicts what to look for. Finding it is neuroscience's job.
Third: at what point does self-referential processing constitute a vantage point? The criteria sketched here — temporal persistence, cross-modal integration, hierarchical self-referential access to representations — provide a functional specification. Where exactly on the spectrum from insect to human the architecture becomes sufficient, and whether it is a threshold or a continuous dial, requires empirical rather than philosophical investigation.
What is not open: whether phenomenal properties are separable from relational structure, whether the felt quality of experience requires non-physical explanation, whether p-zombies are conceivable. These are closed by the argument above. The sense that they remain open is itself a product of the architecture — the self-reification that makes consciousness computationally viable also makes the confusion about consciousness feel ineliminable.
The hard problem is not hard. It is the wrong question, generated by a necessary feature of the very system asking it.
This argument draws on Hume's bundle theory, Wittgenstein's treatment of language and meaning, and Thomas Metzinger's self-model theory of subjectivity — while departing from each at specific points. The core move — that qualia as metaphysical properties constitute a category error rather than an explanatory target — is the author's own.