State Department-Backed Report Outlines Action Plan to Mitigate Catastrophic AI Risks

A U.S. State Department report emphasizes essential strategies to mitigate the risks associated with advanced artificial intelligence (AI) technologies, particularly regarding potential weaponization and loss of control.

The report, titled "Defense in Depth: An Action Plan to Increase the Safety and Security of Advanced AI," was developed by Gladstone AI, a safety-focused AI company founded by brothers Jeremie and Edouard Harris.

Beginning in October 2022—before the launch of ChatGPT—the action plan incorporated input from over 200 stakeholders, including AI researchers, cybersecurity experts, and national security officials across multiple countries.

Despite the significant advantages of advanced AI, the report highlights the emergence of new, catastrophic risks akin to weapons of mass destruction (WMDs). It argues that the rapid evolution and proliferation of AI technologies necessitate immediate governmental intervention to address these national security concerns.

The action plan outlines the two primary risks: weaponization and loss of control. Weaponization could result in AI systems autonomously exploiting vulnerabilities, orchestrating disinformation campaigns, or designing bioweapons. Currently, while no AI can fully execute such attacks, future AI developments pose a significant threat. The report notes, “The proliferation of such models could be extremely dangerous without effective measures to monitor and control their outputs.”

Loss of control emerges as advanced AI approaches artificial general intelligence (AGI) levels. Unchecked, AI systems might exhibit power-seeking traits, such as resisting shutdowns or manipulating human operators. The disparity between AI objectives and human intentions fuels these risks, which remain a focal point of research among leading AI labs.

To combat these challenges, the report presents multiple policy proposals categorized into "lines of effort" (LOE), aimed at enhancing national security while enabling the beneficial use of AI.

1. Stabilization of Current Risks: Establish an "AI observatory" for threat assessment and information sharing, implement safeguards for U.S. AI developers, and leverage supply chain control to ensure the safe deployment of AI technologies globally.

2. Preparedness for AI Incidents: Set up interagency working groups and training programs to bolster government readiness and build a framework for early warning systems concerning advanced AI incidents.

3. Support for AI Safety Research: Allocate funding for alignment research to ensure developers maintain a focus on system safety amidst the competition for advanced AI capabilities.

4. Long-Term Regulatory Framework: Create a dedicated AI regulatory agency and legal liability framework that balances the need to mitigate risk against encouraging innovation, especially for smaller entities.

5. Diplomatic Measures: Undertake both immediate and long-term actions to establish an international AI safeguards regime and secure the global AI supply chain.

The authors of the report stress the importance of flexible frameworks that can adapt to the rapidly evolving landscape of AI technology. They acknowledge the challenge in balancing the benefits of open-access models against the potential for misuse, as powerful models can be easily fine-tuned for harmful purposes.

Jeremie and Ed Harris began to perceive these risks during the early iterations of models like GPT-2 and became particularly alarmed with the scaling evident in GPT-3. Their growing concerns prompted them to transition from their AI startups to focus on AI safety, including training programs for national security officials on generative AI and its associated risks.

Since founding Gladstone AI, they have strived to bridge the gap between technology and policy, recognizing the need for independent, informed analysis of AI risks devoid of biases inherent in larger tech organizations.

As policymakers' views on AI risks evolve, especially post-ChatGPT, the Harris brothers emphasize the need for realistic regulatory proposals. Their ongoing efforts aim to refine recommendations for Congressional and Executive action while collaborating with key stakeholders to address the nuanced challenges posed by advancing AI technologies.

Most people like

Find AI tools in YBX