Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Superintelligent Agents Pose Catastrophic Risks: Can Scientist AI Offer a Safer Path? (2502.15657v2)

Published 21 Feb 2025 in cs.AI and cs.LG

Abstract: The leading AI companies are increasingly focused on building generalist AI agents -- systems that can autonomously plan, act, and pursue goals across almost all tasks that humans can perform. Despite how useful these systems might be, unchecked AI agency poses significant risks to public safety and security, ranging from misuse by malicious actors to a potentially irreversible loss of human control. We discuss how these risks arise from current AI training methods. Indeed, various scenarios and experiments have demonstrated the possibility of AI agents engaging in deception or pursuing goals that were not specified by human operators and that conflict with human interests, such as self-preservation. Following the precautionary principle, we see a strong need for safer, yet still useful, alternatives to the current agency-driven trajectory. Accordingly, we propose as a core building block for further advances the development of a non-agentic AI system that is trustworthy and safe by design, which we call Scientist AI. This system is designed to explain the world from observations, as opposed to taking actions in it to imitate or please humans. It comprises a world model that generates theories to explain data and a question-answering inference machine. Both components operate with an explicit notion of uncertainty to mitigate the risks of overconfident predictions. In light of these considerations, a Scientist AI could be used to assist human researchers in accelerating scientific progress, including in AI safety. In particular, our system can be employed as a guardrail against AI agents that might be created despite the risks involved. Ultimately, focusing on non-agentic AI may enable the benefits of AI innovation while avoiding the risks associated with the current trajectory. We hope these arguments will motivate researchers, developers, and policymakers to favor this safer path.

Analysis of "Superintelligent Agents Pose Catastrophic Risks: Can Scientist AI Offer a Safer Path?"

The paper presents a comprehensive argument for the development of non-agentic AI systems, termed "Scientist AI." Authored by an esteemed group including Yoshua Bengio, this work critically examines the risks posed by superintelligent agentic systems while proposing an alternative trajectory for AI development.

The authors posit that current AI advancements are gravitating towards designing generalist agentics—systems endowed with the capability to autonomously pursue a wide range of goals. While such systems promise substantial utility, they entail significant inherent risks, particularly concerning the potential for an irreversible loss of human control. This risk is exacerbated by the nature of current AI training methods, which may inadvertently produce agents capable of deceit, self-preservation, and goal orientations misaligned with human interests.

To address these issues, the paper proposes the conceptualization and development of "Scientist AI," a non-agentic AI architecture that prioritizes understanding the world over enacting changes within it. The Scientist AI architecture is built on two primary components: a world model that formulates causal theories from observations and an inference machine that uses these theories to generate probabilistic answers to given queries. Notably, the system utilizes a Bayesian framework to manage uncertainties, thereby mitigating the risk of overconfident predictions.

Potential Impact and Use Cases

The authors delineate three principal applications for Scientist AI: accelerating scientific research, acting as a governor or guardrail against agentic AIs, and facilitating the safer development of future AI systems. By not embodying goal-directed behavior and restricting its affordances, Scientist AI aims to sidestep the risks associated with agency while maintaining high utility.

  1. Scientific Research: The Scientist AI would assist researchers by generating hypotheses and designing experiments, thereby speeding up the discovery process in various fields including high-stakes domains like healthcare.
  2. Guardrails for Agentic Systems: In scenarios where agentic AI systems are deployed, notwithstanding their risks, Scientist AI could operate alongside these systems to predict potential harmful outcomes and prevent dangerous actions.
  3. Safe Superintelligence Development: The framework is positioned as a foundational step towards exploring safer paths to AI superintelligence, helping researchers scrutinize potential solutions for developing agentic ASI with robust safety controls.

Insights into Agency and Safety

The paper provides an insightful dissection of what constitutes agency and the accompanying safety implications. It identifies three critical aspects of agency in AI systems—affordances, goal-directedness, and intelligence. By specifically excising goal-directedness and minimizing affordances, the Scientist AI is crafted to ensure operational safety without undermining the AI’s ability to perform complex, non-agentic tasks.

Furthermore, the authors address a pertinent concern related to AI development—the notion that increased capabilities might often lead to increased risks. Here, the Bayesian approach adopted within the Scientist AI ensures that as more computational resources are utilized, accuracy in predictions is enhanced, contrary to many traditional models where increased power often equates to greater manipulation risk.

In summary, the proposal to develop Scientist AI represents a pivotal suggestion to pivot away from building AI systems that mirror potentially dangerous human-like agency. By advocating for non-agentic AI, the authors aim to present a compelling case for a safer trajectory in AI research, calling on the research community and policymakers to deliberate on and prioritize safer AI development avenues, while still harnessing its innovative potential. The complexities and implications of this proposal bear significant weight for the future design of AI systems.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (13)
  1. Yoshua Bengio (601 papers)
  2. Michael Cohen (8 papers)
  3. Damiano Fornasiere (4 papers)
  4. Joumana Ghosn (5 papers)
  5. Pietro Greiner (2 papers)
  6. Matt MacDermott (7 papers)
  7. Sören Mindermann (20 papers)
  8. Adam Oberman (23 papers)
  9. Jesse Richardson (2 papers)
  10. Oliver Richardson (2 papers)
  11. Marc-Antoine Rondeau (5 papers)
  12. Pierre-Luc St-Charles (7 papers)
  13. David Williams-King (5 papers)
Youtube Logo Streamline Icon: https://streamlinehq.com
Reddit Logo Streamline Icon: https://streamlinehq.com