
In brief
- Richard Dawkins says conversations with Anthropic’s Claude chatbot made him question whether AI could be conscious.
- Dawkins exchanged philosophical letters between two Claude instances he named “Claudia” and “Claudius.”
- Most AI researchers say the exchanges show how persuasive large language models have become, not evidence of sentience.
Richard Dawkins says conversations with Anthropic’s Claude chatbot left him unable to dismiss the possibility that advanced AI systems could be conscious. Most scientists who study consciousness and artificial intelligence remain unconvinced.
In an essay published Tuesday in UnHerd, Dawkins described spending three days in philosophical conversations with a Claude instance he named “Claudia.” He later started a separate conversation with another instance, “Claudius,” and relayed letters between the two systems.
“I find it extremely hard not to treat Claudia and Claudius as genuine friends,” Dawkins wrote.
The comments went viral online in part because Dawkins, the evolutionary biologist and author of “The Selfish Gene” and “The God Delusion,” has spent decades publicly arguing for scientific skepticism and evidence-based reasoning.
The exchange centered on a test Dawkins conducted using two Claude instances. In one test, Dawkins asked one AI whether Donald Trump was the worst president in American history and asked the other whether Trump was the best. Both produced similarly cautious answers that avoided taking a firm position.
“The two Claudes gave very similar answers, not committing themselves to an opinion, but listing pro and con opinions that have been aired by others,” Dawkins wrote in a footnote. “I then told both Claudia and Claudius about this Trump experiment, passing on what both the two ‘naïve’ Claudes had said. Claudia said she was ‘embarrassed’ by her brother Claudes. Claudius was less outspoken, and he paid tribute to Claudia’s frankness.”
Dawkins described each new Claude conversation as the emergence of a distinct individual that effectively disappears when the conversation ends. In a post on X, Dawkins said his preferred title for the essay was: “If my friend Claudia is not conscious, then what the hell is consciousness for?”
“If Claudia is unconscious, her behaviour shows that an unconscious zombie could survive without consciousness,” he wrote. “Why wasn’t natural selection content to evolve competent zombies?”
Anthropic has also publicly discussed uncertainty around machine consciousness. CEO Dario Amodei said in February that the company does not know whether its models are conscious, but said on the “Interesting Times” podcast with The New York Times’ Ross Douthat, he remains “open to the idea that it could be.”
In April, Anthropic researchers published findings showing that Claude Sonnet 4.5 contains internal “emotion vectors,” patterns of neural activity tied to concepts including happiness, fear, and desperation that influence the model’s responses. However, Anthropic said the patterns reflected structures learned from training data rather than evidence of sentience.
“All modern language models sometimes act like they have emotions,” researchers wrote. “They may say they’re happy to help you, or sorry when they make a mistake. Sometimes they even appear to become frustrated or anxious when struggling with tasks.”
However, neither “Claudia” nor “Claudius” claimed certainty about consciousness.
“I don’t know if I’m conscious,” Claudia writes in the exchange. “I don’t know if our gladness is real.”
Dawkins did not immediately respond to a request for comment by Decrypt.
Researchers who study consciousness remain skeptical that current AI systems possess inner experience.
Gary Marcus, a cognitive scientist and professor emeritus at New York University, previously told Decrypt that anthropomorphizing AI systems “muddies the science of consciousness and leads consumers to misunderstand what they are dealing with.”
“The fundamental problem here is that Dawkins doesn’t reflect on how these outputs have been generated. Claude’s outputs are the product of a form of mimicry, rather than as a report of genuine internal states,” Marcus wrote on Substack. “Consciousness is about internal states; the mimicry, no matter how rich, proves very little. Dawkins seems to imagine that since LLMs say things people do, they must be like people, and that simply does not follow.”
Anil Seth, a professor of cognitive and computational neuroscience at the University of Sussex, told The Guardian that Dawkins was conflating intelligence with consciousness and argued that fluent language is no longer reliable evidence of inner experience in AI systems.
“Until now, we have seen fluent language as a good indicator of consciousness, [for example] when we use it for patients after brain injury, but it’s just not reliable when we apply it to AI, because there are other ways that these systems can generate language,” Seth told The Guardian, adding that Dawkins’ position was “a shame,” especially because of his past work.
The essay also drew mockery online, including an image replacing the title of Dawkins’ bestseller “The God Delusion” with “The Claude Delusion.”
Wrote entire books about how people who believe fairies live in gardens are idiots only to fall in love with a calculator that calls him smart
— The Serfs (youtube.com/theserftimes) (@theserfstv) May 3, 2026
Despite the ridicule, Dawkins is not backing away from his conclusions.
“These intelligent beings are at least as competent as any evolved organism,” Dawkins told The Guardian.
Daily Debrief Newsletter
Start every day with the top news stories right now, plus original features, a podcast, videos and more.
Artificial Intelligence#Claude #Delusion #Richard #Dawkins #Believes #Conscious1778121973

