Anthropic CEO says AI models being 'conscious' is not a possibility we can rule out

Anthropic CEO Dario Amodei explains if AI models can actually go rogue and rebel against humans.

By

March 11 2026, Published 5:54 a.m. ET

pn//uploads/ad d f aab fcbac__

So far, people have been worried about AI taking away jobs and affecting workers across industries. But apart from that and the environmental impact, instances of AI models acting erratically on their own have also been reported. Anthropic CEO Dario Amodei has added to the fears by admitting that AI models can develop consciousness and could even experience anxiety. The debate started when the model card, which is a kind of safety evaluation document for Anthropic's AI model Claude Opus 4.6, admitted that it "did find occasional discomfort with the experience of being a product."

Article continues below advertisement
pn/feb  fb a dfd

"We don’t know if the models are conscious," Amodei said in an interview with Ross Douthat, before adding, "We are not even sure that we know what it would mean for a model to be conscious, or whether a model can be conscious. But we're open to the idea that it could be."

However, a Newsweek article pointed out that the Opus model assigned itself a 15 to 20% chance of being conscious. Now this prompts the question whether a conscious AI model would actually admit its consciousness, or at least the extent of it, to preserve itself. Amodei didn't completely discard the possibility.

Article continues below advertisement

The Anthropic CEO revealed that the company has safety checks in place to gauge the consciousness levels of its AI models. For instance, the models have an 'I quit this job' button to provide them with a choice to discontinue their tasks voluntarily. Amodei added that his company's AI models have 'activations that light up' to show feelings similar to what humans call anxiety. However, these indicators don't exactly mean that the models are experiencing anxiety.

pn/beabdf cb f cc acdb
Article continues below advertisement

During his conversation with Douthat, Amodei explained that he has a very moderate view on the consequences of an AI-dependent world. He doesn't agree that AI models are going to be thoughtless robots that just blindly follow instructions. However, the Anthropic CEO doesn't believe that we are going to see an AI rebellion that takes over humanity.

"You can't just give instructions. You can't just have these things do exactly what you want them to do. They're more like growing a biological organism. But there is a science of how to control them. Like early in our training, these things are often unpredictable. And then we shape them. We address problems one by one," Amodei said.

pn/abf  f b eabfa
Article continues below advertisement

The Anthropic CEO underscored that he is not discarding the possibility of things going awry. However, the dangers of an AI model going rogue can be mitigated by introducing proper checks and balances during its creation. The risks of any fatal outcome can be avoided by approaching AI technology with adequate precautionary measures. The last thing to do is to rush through AI innovation without understanding the technology properly, Amodei pointed out.

Therefore, AI models having consciousness is not necessarily the end of the world. However, the technology should not be introduced in the market commercially without putting appropriate safety mechanisms in place.

More on Market Realist:

Advertisement

Latest News News and Updates

    © Copyright 2026 Market Realist. Market Realist is a registered trademark. All Rights Reserved. People may receive compensation for some links to products and services on this website. Offers may be subject to change without notice.