Artificial Intelligence & Machine Learning
,
Next-Generation Technologies & Secure Development
Enthusiasm for AI Development Is Outpacing Discussions on Safety

Google DeepMind executives outlined an approach to artificial general intelligence safety, warning of “severe harm” that can “permanently destroy humanity” if safeguards are not put in place before advanced artificial intelligence systems emerge.
See Also: Capturing the cybersecurity dividend
A 145-page paper forecasts that AGI could arrive by 2030, potentially capable of performing at the 99th percentile of skilled adults in a wide range of non-physical tasks. The company called for proactive risk mitigation strategies as competitive pressures drive AI development.
The paper identified four major areas of concern: deliberate misuse, misalignment between AI actions and human intent, accidental harm, and structural risks arising from AI system interactions.
Paper authors Anca Dragan, Rohin Shah, Four Flynn and Shane Legg proposed a mix of technical and policy interventions to address these challenges, focusing on training, monitoring and security. A key discussion point of the paper is whether AGI could lead to recursive AI improvement, where AI systems conduct their own research to enhance future models. The authors said that such a feedback loop could pose serious risks.
But some experts are skeptical. AI researcher Matthew Guzdial reportedly dismissed the idea as speculative, noting a lack of evidence supporting self-improving AI systems. AI regulation expert Sandra Wachter told TechCrunch that the focus must be on a more immediate issue: AI systems learning from their own flawed outputs, reinforcing inaccuracies over time.
DeepMind’s concerns come at a time when enthusiasm for AI development is outpacing discussions on safety. Global competition, particularly between the United States and China, is accelerating the race to AGI. U.S. Vice President JD Vance dismissed excessive caution at the Paris AI Action Summit, arguing that AI progress depends on building infrastructure rather than debating hypothetical dangers. Google CEO Sundar Pichai reinforced this sentiment, saying AI has the potential to drive positive change despite historical fears surrounding new technologies.
Some AI researchers challenge this optimism. AI pioneer Yoshua Bengio criticized the Paris AI Summit’s lack of urgency on safety, warning that AI risks demand more serious attention. Anthropic CEO Dario Amodei echoed the concerns, advocating for increased focus on AI safety as the technology advances rapidly.
Industry players do agree that today’s AI systems already exhibit unexpected behaviors. A recent study by Anthropic found that large language models demonstrate advanced reasoning capabilities beyond what their creators anticipated. It observed instances where AI systems planned steps ahead to compose poetry, challenging prior assumptions about their cognitive processes. Cases of AI models finding workarounds for missing computational resources have also emerged, illustrating the potential for unintended consequences (see: A Peek Into How AI ‘Thinks’ – and Why It Hallucinates).
The DeepMind paper does not provide definitive solutions but looks to guide discussions on AI risk mitigation. Authors advised continued research into AI safety, better understanding of AI decision-making and stronger protections against malicious use.
“The transformative nature of AGI has the potential for both incredible benefits as well as severe harms,” the DeepMind authors wrote. “As a result, to build AGI responsibly, it is critical for frontier AI developers to proactively plan to mitigate severe harms.”