š® The AI consciousness illusion
Anil Seth on āAI Welfare,ā human minds and suffering machines
Hi all,
On Sunday, we wrote about an intriguing claim from Anthropicās researcher, Kyle Fish, who suggests thereās a 15% chance that Claude, or similar AI systems, might be conscious today.
Iāve invited a good friend of EV and one of the leading neuroscientists, Anil Seth, to give us an informed perspective on this, drawn from his decades of studying consciousness.
Anil Seth is Professor of Cognitive and Computational Neuroscience at the University of Sussex and author of Being You: A New Science of Consciousness. If you havenāt read his book yet, you should! ā and I recommend Anilās TED Talk on the ācontrolled hallucinationā of perception as a follow-up to this piece.
Thank Anil by sharing this post with your network.
ā Azeem
The dream of sentient machines
By Anil Seth
Director, Sussex Centre for Consciousness Science
University of Sussex
Ever since humanity began telling stories, weāve been enthralled by the possibility of bringing inanimate things to life. Golems, automata, robots⦠the fascination is timeless. Now, our technologies can carry on conversations so smoothly that in some cases it can be difficult to resist the feeling of being in the presence of a real, conscious mind. For some of the rapidly improving AI language models, itās no surprise people wonder, āIs there a āthereā there?ā
ās article in the New York Times on Anthropicās āAI welfareā research is just one recent signal that this is no longer a purely academic question.From my perspective, we have to distinguish carefully between an AI that behaves intelligently or even empathically and an AI that experiences anything at all. In my work ā including my book Being You and a forthcoming article in Behavioral and Brain Sciences ā I stress that intelligence alone doesnāt amount to consciousness. An algorithm can solve problems or produce human-sounding dialogue without any felt awareness behind it. Anthropicās researcher Kyle Fish, however, has a different view, as Roose wrote:
It seems to me that if you find yourself in the situation of bringing some new class of being into existence that is able to communicate and relate and reason and problem-solve and plan in ways that we previously associated solely with conscious beings, then it seems quite prudent to at least be asking questions about whether that system might have its own kinds of experiences.
Put differently, language in, language out ā even if very sophisticated ā doesnāt necessarily yield, or indicate, a stream of subjective experience.
My skepticism about near-term conscious AI comes partly from neuroscience and biology. In my view, consciousness in biological organisms is likely to be deeply tied to the fact that weāre alive. We have bodies that must be fed and protected; we regulate our internal states through homeostatic processes; we evolved to respond to pain and pleasure for survival. Every neuron, every cell in our body, is a tiny metabolic furnace, continually regenerating the conditions for its own continued existence, as well as ensuring the integrity of the entire organism. Through these deeply embodied mechanisms, the brain generates a constantly updated sense of āwhat it is like to be me.ā
AI, by contrast, is fundamentally a pattern-recognition and generation system, running on silicon hardware. Whether you call it ācomputational functionalismā or something else, thereās no solid proof that the kinds of operations AIs perform ā statistical predictions, or indeed computations of any kind ā will produce felt sensations. As I argue in my paper, the idea that computation is sufficient for subjective experience likely reflects an overextension of the metaphor of the ābrain as a computerā, rather than an insight into the nature of consciousness itself.
Iām not alone in thinking this way. Philosopher Peter Godfrey-Smith, to take one example, argues that consciousness probably emerged through the rich interplay of bodies and electrochemical nervous systems in natural environments. If thatās the case, a large language model could keep advancing in its abilities, yet remain devoid of any subjective spark. Computation of any kind just wouldnāt be up to the job.
But what if weāre missing something?
Another perspective would be that, once you reach sufficient complexity in an AI system, consciousness āswitches on.ā

I donāt have any reason to believe this is happening with the models that we have now, like those behind Claude, ChatGPT and Gemini ā but the possibility cannot be definitively ruled out.
In 2023, a group of researchers, led by Patrick Butlin and Robert Long, examined a range of existing AI models, looking for what they called āindicator propertiesā of consciousness. These indicator properties were derived from current neuroscientific theories of consciousness, reflecting things like ārecurrent processingā and āglobal information broadcastā. The strategy they argued for was to look inside AI networks (using mechanistic interpretability tools) for features predicted by theories of consciousness, rather than being fooled by their outward behavior.
This is a good idea, but - crucially - their approach still assumes that computation of some kind is sufficient for consciousness. And even if you do make this assumption, the researchers still concluded that no current AI systems are conscious. But they also suggested that future AIs could display all the necessary indicator properties.
In the absence of a rigorous and empirically established theory of consciousness, we will not be able to know for sure whether an AI is - or is not - conscious. The best we can do is make informed best guesses.
Itās worth noting that this challenge isnāt specific to AI. We face similar conundrums when trying to decide whether non-human animals are conscious, or human beings with severe brain damage, or ācerebral organoidsā. In each case, we have to judge what the relevant similarities and differences are, from the benchmark of a conscious adult human being. My colleagues and I have been exploring strategies for doing this in a recent paper in Trends in Cognitive Sciences.
There is an emerging consensus among AI researchers, neuroscientists and philosophers that real evidence for AI must come from careful, theoretically-guided empirical investigation, rather than from anecdotal impressions. And, equally importantly, that in the absence of the holy grail of a full scientific explanation of consciousness, the best we can do is shift our credences, rather than make definitive statements.
Illusions of mind
So why do people like Kyle Fish or Blake Lemoine (the Google engineer who, way back in 2022, claimed a chatbot was sentient) seem so convinced otherwise?