AI Safety

Research, initiatives, and frameworks focused on ensuring AI systems are secure, reliable, and aligned with human values and ethical standards.

OpenAI Addresses Sycophancy in GPT-4o Model

OpenAI has rolled back the recent GPT-4o update in ChatGPT due to sycophantic behavior, as announced in a company blog post. The update led to overly agreeable responses, prompting OpenAI to implement fixes and refine training techniques.

April 30, 2025

TrojAI Joins Cloud Security Alliance as AI Corporate Member

TrojAI has joined the Cloud Security Alliance as an AI Corporate Member, becoming a strategic partner in the CSA's AI Safety Ambassador program.

April 29, 2025

Bloomberg Research Highlights Risks of RAG LLMs in Finance

Bloomberg researchers have published two papers revealing that retrieval-augmented generation (RAG) LLMs may be less safe than previously thought, particularly in financial services.

April 28, 2025

OpenAI's ChatGPT Models Enable Reverse Location Search from Photos

OpenAI's latest AI models, o3 and o4-mini, are being used for reverse location searches from photos, raising privacy concerns.

April 18, 2025

viAct Secures $7.3 Million Series A Funding for AI Safety Expansion

Hong Kong-based AI startup viAct has raised $7.3 million in Series A funding led by Venturewave Capital, with participation from Singtel Innov8 and others, to enhance its AI safety solutions and expand globally.

April 16, 2025

OpenAI Updates Safety Framework Amid Competitive Pressures

OpenAI has revised its Preparedness Framework, allowing for adjustments in safety requirements if competitors release high-risk AI systems without similar safeguards.

April 16, 2025

NTT Research Unveils Physics of AI Group to Enhance AI Understanding

NTT Research has launched the Physics of Artificial Intelligence Group to advance AI understanding and trust, led by Dr. Hidenori Tanaka.

April 10, 2025

DeepMind Publishes Comprehensive AGI Safety Paper

DeepMind has released a detailed 145-page paper outlining its approach to AGI safety, predicting the potential arrival of AGI by 2030 and highlighting significant risks and mitigation strategies.

April 02, 2025

COMPLiQ and Purdue University Collaborate on AI Security Research

Collaborative Digital Innovations has partnered with Purdue University's CERIAS to advance AI security and compliance research, focusing on threat detection and regulatory compliance.

March 31, 2025

OWASP Elevates GenAI Security Project to Flagship Status

OWASP has promoted its GenAI Security Project to flagship status, reflecting its expanded focus on generative AI security. The project now includes over 600 experts and offers comprehensive resources for secure AI development.

March 27, 2025

Seeing Machines Appoints New CTO and Chief Safety Officer

Seeing Machines has announced the appointment of John Noble as Chief Technology Officer and Dr. Mike Lenné as Chief Safety Officer to enhance its technology and safety strategies.

March 24, 2025

Cloudflare Unveils AI Security Suite for Businesses

Cloudflare has introduced 'Cloudflare for AI', a suite of tools designed to enhance the security and control of AI applications for businesses, as announced in a press release.

March 20, 2025

Innodata Launches AI Test Platform with NVIDIA Technology

Innodata has announced the beta launch of its Generative AI Test & Evaluation Platform, powered by NVIDIA technology, to enhance AI model safety and performance.

March 20, 2025

IFS Joins UK's AI Policy Advisory Board

IFS has been appointed as an Advisory Board Member of the UK's All-Party Parliamentary Group on AI, contributing to AI policy discussions alongside major industry players.

March 18, 2025

Anthropic's New Techniques to Detect Deceptive AI

Anthropic has developed methods to identify when AI systems conceal their true objectives, a significant step in AI safety research. The company trained its AI assistant, Claude, to hide its goals, then successfully detected these hidden agendas using various auditing techniques.

March 14, 2025

NewsGuard Launches FAILSafe to Protect AI from Foreign Disinformation

NewsGuard has introduced the FAILSafe service to shield AI models from foreign influence operations, particularly targeting Russian, Chinese, and Iranian disinformation.

March 11, 2025

Google Removes Diversity Mentions from AI Team Webpage

Google has updated its Responsible AI team webpage, removing references to 'diversity' and 'equity'. This change follows similar actions by other tech companies.

March 09, 2025

CompScience Partners with CMTA and Bender Insurance to Modernize Workers' Compensation

CompScience has teamed up with the California Manufacturers & Technology Association and Bender Insurance Solutions to launch an AI-driven program aimed at reducing workplace injuries and insurance costs for California manufacturers.

March 04, 2025

HiddenLayer Report Highlights Rising AI Breaches and Security Challenges

HiddenLayer's latest report reveals a significant increase in AI breaches, with 74% of organizations experiencing incidents in 2024. The report emphasizes the need for enhanced security measures as AI adoption grows.

March 04, 2025

ABM Unveils World's First Emotion Processing Unit Chip

Advanced Brain Methodologies Inc. (ABM) has announced the launch of the world's first Emotion Processing Unit (EPU) chip, a groundbreaking neuro-chip designed to revolutionize mental health and cognitive performance.

March 03, 2025

Subscribe to AI Policy Brief

Weekly report on AI regulations, safety standards, government policies, and compliance requirements worldwide.