AI Safety

Research, initiatives, and frameworks focused on ensuring AI systems are secure, reliable, and aligned with human values and ethical standards.

DeepMind Publishes Comprehensive AGI Safety Paper

DeepMind has released a detailed 145-page paper outlining its approach to AGI safety, predicting the potential arrival of AGI by 2030 and highlighting significant risks and mitigation strategies.

April 02, 2025

OWASP Elevates GenAI Security Project to Flagship Status

OWASP has promoted its GenAI Security Project to flagship status, reflecting its expanded focus on generative AI security. The project now includes over 600 experts and offers comprehensive resources for secure AI development.

March 27, 2025

IFS Joins UK's AI Policy Advisory Board

IFS has been appointed as an Advisory Board Member of the UK's All-Party Parliamentary Group on AI, contributing to AI policy discussions alongside major industry players.

March 18, 2025

Anthropic's New Techniques to Detect Deceptive AI

Anthropic has developed methods to identify when AI systems conceal their true objectives, a significant step in AI safety research. The company trained its AI assistant, Claude, to hide its goals, then successfully detected these hidden agendas using various auditing techniques.

March 14, 2025

Subscribe to AI Policy Brief

Weekly report on AI regulations, safety standards, government policies, and compliance requirements worldwide.