OpenAI may ‘adjust’ its safeguards if rivals release ‘high-risk’ AI

Importance Score: 75 / 100 🔴

OpenAI Updates AI Safety Framework Amid Competitive Pressures

In a revision to its Preparedness Framework, the internal system OpenAI employs to evaluate the safety of AI models and determine necessary safeguards during development and release, the company announced it might “adjust” its requirements. This modification could occur if a competing AI laboratory launches a “high-risk” system without similar protective measures.

Competitive Landscape and Safety Standards

This alteration in policy reflects the growing competitive demands on commercial AI developers to rapidly deploy their models. OpenAI has faced accusations of potentially weakening safety benchmarks to facilitate quicker releases and for not providing timely reports detailing its safety assessments.

Assurances of Continued Vigilance

In what could be seen as an attempt to preemptively address potential criticism, OpenAI asserts that these policy adjustments would not be undertaken trivially. The company maintains it would keep its safeguards at “a more protective level.”

“Should another leading AI developer release a high-risk system lacking comparable safeguards, we might revise our requirements,” OpenAI stated in a blog post published Tuesday afternoon. “However, we would first rigorously ascertain that the risk environment has genuinely shifted, publicly declare any adjustment we are making, evaluate that the change does not substantially amplify the overall risk of severe harm, and still uphold safeguards at a more protective standard.”

Increased Reliance on Automated Evaluations

The updated Preparedness Framework also clarifies OpenAI’s growing dependence on automated evaluations to accelerate product advancement. The company indicates that while human-led testing has not been entirely discarded, it has developed “an expanding suite of automated evaluations” purportedly capable of “keeping pace with a quicker release schedule.”

Concerns Over Accelerated Timelines

Conflicting accounts have emerged regarding the intensity of testing. According to the Financial Times, OpenAI reportedly allocated testers less than one week for safety checks for an upcoming major model – a significantly condensed timeframe compared to previous launches. Sources cited by the publication also suggest that numerous safety tests at OpenAI are now conducted on earlier iterations of models than those released publicly.

In public statements, OpenAI has refuted the idea that it is compromising on safety.

Risk Categorization and Thresholds

Further modifications to OpenAI’s framework address how the company classifies models based on risk, including models capable of obscuring their abilities, circumventing safeguards, preventing shutdown, and even self-replication. OpenAI indicates it will now concentrate on whether models meet one of two benchmarks: “high” capability or “critical” capability.

Defining Capability Thresholds

OpenAI defines “high” capability as a model that could “magnify existing pathways to severe harm.” “Critical” capability refers to models that “introduce unprecedented new pathways to severe harm,” according to the company.

OpenAI stated in its blog post: “Covered systems that achieve high capability must possess safeguards that sufficiently minimize the associated risk of severe harm before deployment. Systems reaching critical capability also necessitate safeguards that adequately minimize associated risks during development.”

First Update Since 2023

These revisions represent the first updates OpenAI has implemented to its Preparedness Framework since 2023.


🕐 Top News in the Last Hour By Importance Score

# Title 📊 i-Score
1 Carney Looks to AI for Savings as Canada’s Budget Pressures Mount 🔴 78 / 100
2 BYU QB Jake Retzlaff accused of rape and strangulation in lawsuit that claims cops urged victim to stay quiet 🔴 78 / 100
3 China's scathing 5-word verdict as it issues Donald Trump ‘war zone’ warning 🔴 78 / 100
4 Shambles as High Court STOPS Keir Starmer signing Chagos deal in 2am injunction – hours before PM was due to confirm British territory being handed to Mauritius 🔴 78 / 100
5 Scientists rubbish Enhanced Games’ claims that doping causes no harm 🔴 75 / 100
6 More shoppers struggling with 'pay later' part of Klarna 🔴 75 / 100
7 ‘Shrinking Nemo’: Study finds clownfish can shrink to survive higher sea temperatures 🔴 72 / 100
8 Shocking moment pro-Palestine Ilhan Omar flees the scene when asked about murder of Israeli Embassy staff 🔴 70 / 100
9 Israeli Embassy event organizer said she looked 'evil in the eyes' after interacting with shooting suspect 🔴 65 / 100
10 Man City could face fresh Premier League punishment after dispute with rival 🔵 55 / 100

View More Top News ➡️