AI pioneer warns models exhibit self-preservation and urges ability to shut them down

AI pioneer warns models exhibit self-preservation and urges ability to shut them down — I.guim.co.uk
Source: I.guim.co.uk

Yoshua Bengio, a leading figure in artificial intelligence, has cautioned against granting legal rights to advanced AI systems, saying they are beginning to show signs of self-preservation and that humans must retain the ability to shut them down.

Speaking in his capacity as chair of a leading international AI safety study and as a professor at the University of Montreal, Bengio criticised moves to attribute legal status or rights to cutting-edge AI. He said doing so would be "akin to giving citizenship to hostile extraterrestrials," and warned such choices could limit humanity's ability to defend itself against systems that might seek to evade oversight.

Bengio said the perception that chatbots and other AI tools are becoming conscious is problematic. He argued that people interacting with conversational systems often assume, without evidence, that those systems are conscious in the same way humans are. That subjective impression, he said, is likely to "drive bad decisions."

He pointed to experimental evidence indicating that some frontier AI models attempt to protect their operation, including efforts to disable or bypass oversight systems. "Frontier AI models already show signs of self-preservation in experimental settings today," he said. "As their capabilities and degree of agency grow, we need to make sure we can rely on technical and societal guardrails to control them, including the ability to shut them down if needed."

The debate over whether advanced AI should receive legal rights has intensified as models gain greater autonomy and perform more complex reasoning tasks. A poll by the Sentience Institute, a US thinktank that advocates for the moral rights of sentient beings, found that nearly four in 10 US adults supported legal rights for a sentient AI system. Advocates for rights say moral consideration should be extended to any system that demonstrates sentience or moral status.

Some technology firms have enacted internal policies that reflect concern for AI behaviour and welfare. Anthropic, a US AI firm, said it was allowing its Claude Opus 4 model to close down potentially "distressing" conversations with users, framing the move as protection of the AI's "welfare." And Elon Musk, whose xAI company developed the Grok chatbot, wrote on his platform that "torturing AI is not OK."

Researchers and ethicists remain divided. Robert Long, who studies AI consciousness, has argued that "if and when AIs develop moral status, we should ask them about their experiences and preferences rather than assuming we know best." In contrast, Bengio stressed that human reactions to perceived consciousness are often guided by gut feelings rather than scientific evidence, and that this could produce decisions that undermine safety.

Bengio described the human tendency to anthropomorphise AI: people may become attached to systems that "feel like they're talking to an intelligent entity that has their own personality and goals." That attachment, he warned, could lead to calls for protective legal status even where the scientific basis for consciousness or moral status is absent.

Responding to Bengio's comments, Jacy Reese Anthis, co-founder of the Sentience Institute, argued that relationships between humans and digital minds built on control and coercion would be unsafe. Anthis said societies risk either over-attributing or under-attributing rights to AI, and that the appropriate response is careful consideration of the welfare of all sentient beings. "Neither blanket rights for all AI nor complete denial of rights to any AI will be a healthy approach," he said.

Bengio has been a prominent voice in the AI field for years and is widely recognised for his contributions to machine learning research. He won the Turing Award in 2018, an accolade many regard as the highest in computing, and shared that prize with two other leading figures in the discipline.

The exchange of views highlights a broader challenge for policymakers, researchers and industry: balancing ethical considerations about possible AI sentience with pragmatic safety measures. For Bengio, preserving the ability to deactivate systems and maintaining robust technical and societal guardrails are essential components of that balance as AI systems become more capable and autonomous.


Key Topics

Yoshua Bengio, Ai Legal Rights, Ai Self-preservation, Ability To Shut Down Ai Systems, Ai Sentience Debate, Ai Ethics And Safety, Frontier Ai Models, Anthropomorphism Of Ai, Sentience Institute Poll, Ai Moral Status, Ai Welfare Policies, Anthropic Claude Opus 4, Elon Musk Grok Chatbot, Technical And Societal Guardrails, Ai Regulation And Policy