Safe AI Forum
banner
safe-ai-forum.bsky.social
Safe AI Forum
@safe-ai-forum.bsky.social
Trying to foster international cooperation on AI safety. We run
the International Dialogues on AI Safety (IDAIS). Find us at http://saif.org
The event led to the impactful release of The Singapore Consensus on Global AI Safety Research Priorities, guiding crucial international collaboration. We’re proud to have contributed and look forward to continuing this important work. Read more: aisafetypriorities.org
The Singapore Consensus on Global AI Safety Research Priorities
Building a Trustworthy, Reliable and Secure AI Ecosystem. Read the full report online, or download the PDF.
aisafetypriorities.org
May 8, 2025 at 6:30 PM
Read the full paper here: saif.org/research/bar... or on arxiv: arxiv.org/abs/2504.15416
Bare Minimum Mitigations for Autonomous AI Development - Safe AI Forum
saif.org
May 7, 2025 at 4:53 PM
We propose 4 bare minimum safeguards applicable when AI agents significantly automate or accelerate AI development:
May 7, 2025 at 4:53 PM
2️⃣ AI R&D agents could dramatically accelerate improvement in catastrophic capabilities, leading to adaptation lag for both technical and governance measures as well as dangerous capability proliferation.
May 7, 2025 at 4:53 PM
We identify two core risk pathways:

1️⃣ As AI agents automate AI research workflows, they could sabotage safety efforts and create an unauthorized internal deployment for dangerous purposes.
May 7, 2025 at 4:53 PM
In our latest paper, we outline how autonomous AI R&D could lead to serious risks. This work reflects a vital international dialogue, uniting experts across jurisdictions – including authors from China, the US, Canada & Europe – to address these shared global challenges.
May 7, 2025 at 4:53 PM
Over time, SAIF aims to improve the odds of international AI cooperation. Some other ways we hope to push this is by making policy recommendations, conducting research, and directly facilitating international cooperation ourselves through further convenings of researchers.
May 1, 2025 at 9:50 AM
To date we've hosted three dialogues, in Oxford, Beijing and Venice. (You can read more about them at idais.ai)
May 1, 2025 at 9:50 AM
Our foundational program is the
International Dialogues on AI Safety (IDAIS). Output of these dialogues demonstrates consistent scientific consensus on AI risks, and the very existence of these high-profile China/West meetings demonstrates that meaningful international cooperation is possible.
May 1, 2025 at 9:46 AM