Abstract
As AI capabilities advance, practical steps to ensure safety become critical. Drawing on consensus built through the International Dialogues on AI Safety (IDAIS) in Venice, Beijing, and Oxford, we've developed a comprehensive guide mapping potential actions for different stakeholders.
Policy Guide
Starting in late 2023, the International Dialogues on AI Safety has convened leading scientists and academics from around the world to build consensus on risks from frontier AI systems and the governance interventions needed to tackle these risks. Based on this consensus, IDAIS participants have signed public statements outlining a set of goals for AI safety and governance.
This policy guide aims to connect the goals recommended in the consensus statements to direct policy action that policymakers, philanthropists, companies, and researchers can consider taking to improve the state of AI safety and governance.
The guide is split into four key policy areas, based on the goals recommended in the statements:
-
AI safety research
-
Testing and evaluation
-
Domestic governance
-
International governance
For each area, we examine specific challenges - from verification methods to monitoring systems - and outline potential policy levers, drawing on both historical examples and emerging practices.
This is a living document that will evolve with future dialogues and emerging governance challenges. It aims to serve as a practical resource for policymakers, companies, researchers, and philanthropists working to advance AI safety.