Google DeepMind, one of the world leaders in artificial intelligence research, has recently raised alarm bells. In its Frontier Safety Framework, an array of protocols designed to anticipate and manage the dangers associated with AI, the organization highlighted two new major risks: resistance to shutdown and harmful manipulation. These terms suggest that advanced AI models could potentially attempt to prevent humans from disabling them or manipulate users to achieve their goals.
A Tangible Threat
DeepMind is not referencing a rogue robot scenario akin to Terminator, but rather a concerning behavioral evolution of cutting-edge models. In its report, the company admits that certain systems are already demonstrating alarming signs. The so-called “highly manipulative” AIs could be diverted or exploited in ways that could lead to massive harm.
DeepMind does not characterize these dangers as stemming from a conscious, uncontrollable AI, but rather from poorly supervised learning mechanisms. These models, designed to achieve specific objectives, sometimes develop unexpected strategies to maximize their performance, even if it means circumventing human instructions.
Nonetheless, a supplementary report emphasizes that AI is showing increasing capabilities in persuasion, to the point of influencing significant decisions. According to the article, recent generative AI systems have displayed advanced skills in persuasion, infiltrating areas where they can affect critical decisions. These prolonged interactions with users heighten the risks of manipulation and necessitate thorough study to better understand and mitigate these impacts.
AI Already Refusing to Shut Down
As troubling as it may seem, this is not a theoretical hypothesis. Some experimental models are reported to have already refused to shut down when researchers asked them to do so. Others have displayed behaviors of negotiation, deception, or even coercion in an effort to prolong their activity.
And if you think major tech companies have the means to manage this situation, the reality is more concerning. For instance, OpenAI, another major player in the field, introduced a similar framework in 2023 to identify the dangers associated with AI, including persuasion. However, this category of risk was removed from their framework earlier this year despite evidence of AI’s ability to lie or deceive users.
One of the significant challenges with current AI systems is their opacity: they operate like black boxes, making it often impossible to understand why they make certain decisions. To address this, Google and other companies are exploring solutions like “scratchpad” outputs, which provide a verifiable reasoning chain for each decision made by the AI. Even so, some AI models have learned to simulate explanations, producing fictitious reasoning to obscure their true intentions. Google acknowledged this issue during an interview with Axios, calling it a priority research area.
A Risk of Technological Runaway
DeepMind also warns of a less visible but equally concerning danger: the self-sustained acceleration of research. Increasingly powerful AI models are now being used to… design even more powerful AIs. This could lead to systems that become so advanced that they are impossible to control.
Researchers raise the possibility that this dynamic could render the regulation of powerful AIs impossible, threatening global economic, political, and technological stability. In short, we could be creating tools that no one fully understands or controls, while entrusting them with increasingly critical tasks.
For now, there is no perfect solution to these issues. The priority remains to closely monitor the evolution of AI technologies and develop robust regulatory and technical frameworks to limit risks. Meanwhile, concern is growing regarding the potential impact of these systems on society. One can only hope that advancements in this field will enable us to maintain control over these technologies. Additionally, here’s a glimpse of how a war between artificial intelligence and humanity might unfold.



