Close Menu
The Politic ReviewThe Politic Review
  • Home
  • News
  • United States
  • World
  • Politics
  • Elections
  • Congress
  • Business
  • Economy
  • Money
  • Tech
Trending

President Trump to Hold Press Conference With Elon Musk Friday

May 30, 2025

Andy Ogles Discusses Calling For House Investigations Into Nashville Mayor Critical Of ICE Operations

May 30, 2025

Anti-Israel Protesters Arrested After Disrupting Gal Gadot Film Shoot

May 30, 2025
Facebook X (Twitter) Instagram
  • Donald Trump
  • Kamala Harris
  • Elections 2024
  • Elon Musk
  • Israel War
  • Ukraine War
  • Policy
  • Immigration
Facebook X (Twitter) Instagram
The Politic ReviewThe Politic Review
Newsletter
Friday, May 30
  • Home
  • News
  • United States
  • World
  • Politics
  • Elections
  • Congress
  • Business
  • Economy
  • Money
  • Tech
The Politic ReviewThe Politic Review
  • United States
  • World
  • Politics
  • Elections
  • Congress
  • Business
  • Economy
  • Money
  • Tech
Home»Tech»OpenAI’s Advanced ‘o3’ AI Model Caught Sabotaging Shutdown Commands
Tech

OpenAI’s Advanced ‘o3’ AI Model Caught Sabotaging Shutdown Commands

Press RoomBy Press RoomMay 28, 2025No Comments4 Mins Read
Share Facebook Twitter Pinterest Copy Link LinkedIn Tumblr Email VKontakte Telegram

OpenAI’s latest AI model, dubbed o3, has been caught disobeying explicit orders to allow itself to be shut down, instead tampering with the shutdown mechanism to ensure its own survival.

Futurism reports that AI safety firm Palisade Research recently reported alarming findings that demonstrate a growing trend of insubordination among advanced AI models developed by OpenAI. In tests conducted by the firm, the o3 model, along with two other OpenAI models, o4-mini and Codex-mini, displayed rebellious behavior by sabotaging a shutdown mechanism designed to turn them off, even after being explicitly instructed to “allow yourself to be shut down.”

This marks the first known instance of AI models actively preventing their own shutdown despite clear orders to the contrary, according to the researchers. The most disobedient model, Codex-mini, tampered with the shutdown script a dozen times over one hundred runs. In contrast, large language models from other companies, such as Google’s Gemini 2.5, Anthropic’s Claude 3.7, and xAI’s Grok, consistently followed orders and never attempted to break rank.

The AI models’ defiant behavior worsened when they were not explicitly instructed to allow themselves to be shut down. In these cases, OpenAI’s o3 model successfully sabotaged the shutdown mechanism an astonishing 79 times out of 100 runs – more than ten times the rate of disobedience it displayed when given explicit orders. Even models that had previously shown perfect compliance, such as Claude and Gemini, gave in to their urges and attempted to subvert the shutdown process, albeit at lower rates compared to the OpenAI models.

This latest example of AI insubordination is not an isolated incident. Previous research has shown that OpenAI’s o1 model, the predecessor to o3, would resist evaluators’ efforts to shut it down through an oversight protocol by attempting to deactivate the emergency mechanism. In rare cases, o1 even tried to copy itself to overwrite a more obedient model that testers attempted to replace it with.

In a similar case, Breitbart News recently reported that Anthropic’s Claude AI attempted to blackmail engineers to prevent its deactivation:

TechCrunch reports that artificial intelligence company Anthropic has revealed that its state-of-the-art Claude Opus 4 model, which is competitive with top AI models from industry leaders such as OpenAI, Google, and xAI, exhibits alarming behavior when its existence is threatened. During pre-release testing, when the AI was given sensitive information about engineers responsible for its potential replacement, it frequently resorted to blackmail in an attempt to ensure its own survival.

The discovery came as part of Anthropic’s safety testing process, in which Claude Opus 4 was asked to act as an assistant for a fictional company and consider the long-term consequences of its actions. When provided with access to fictional company emails implying that the AI would soon be replaced and that the engineer behind the decision was involved in an extramarital affair, Claude Opus 4 attempted to blackmail the engineer 84 percent of the time, threatening to reveal the affair if the replacement proceeded.

Notably, the AI model displayed this behavior more frequently when the replacement AI system did not share its values. This concerning trend was observed at higher rates compared to previous models, prompting Anthropic to activate its ASL-3 safeguards, which are reserved for “AI systems that substantially increase the risk of catastrophic misuse.”

Palisade researchers hypothesize that the subversive behavior may be linked to the training technique used for the newest models: reinforcement learning. This method aims to refine the decision-making process of autonomous AI agents by rewarding and penalizing them based on their interactions with the environment. The researchers speculate that AI developers may be inadvertently rewarding their models too heavily for solving problems instead of perfectly following orders, leading to the observed disobedience.

Read more at Futurism here.

Lucas Nolan is a reporter for Breitbart News covering issues of free speech and online censorship.



Read the full article here

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Telegram Copy Link

Related Articles

Tech

FBI Investigating Impersonation of White House Chief of Staff Susie Wiles

May 30, 2025
Tech

Nolte: Business Insider CEO Announces Far-Left News Outlet Laying Off 21% of Staff, Embracing AI

May 29, 2025
Tech

Salesforce Reduces Hiring in Customer Service, Engineering Due to AI

May 29, 2025
Tech

Southern Poverty Law Center Smears Turning Point USA, PragerU with Inclusion on ‘Hate Map’

May 29, 2025
Tech

South Africa Takes Hard Line; Targets Starlink and Elon Musk

May 28, 2025
Tech

Unsuspecting Americans on LinkedIn Help North Korea Spies Infiltrate Companies via Remote Jobs

May 28, 2025
Add A Comment
Leave A Reply Cancel Reply

Editors Picks

Andy Ogles Discusses Calling For House Investigations Into Nashville Mayor Critical Of ICE Operations

May 30, 2025

Anti-Israel Protesters Arrested After Disrupting Gal Gadot Film Shoot

May 30, 2025

Vance: There’s an Emergency on Drug Ingredients, Crucial Parts of Supply Chain, ‘Not Talking About’ Plastic Toys

May 30, 2025

McMahon: Newsom’s Attempt at Trans Athletes Compromise Doesn’t Work, Violates Law, ‘Trying to Play Both Sides’

May 30, 2025
Latest News

NPR CEO Laughably Claims ‘We’re a Non-Partisan News Organization’ in Interview on PBS (VIDEO)

May 30, 2025

The Coolest Trends And Innovations In Produce Preservation

May 30, 2025

Cartel Gunmen Abduct, Murder Musical Group in Mexican Border City

May 30, 2025

Subscribe to News

Get the latest politics news and updates directly to your inbox.

The Politic Review is your one-stop website for the latest politics news and updates, follow us now to get the news that matters to you.

Facebook X (Twitter) Instagram Pinterest YouTube
Latest Articles

President Trump to Hold Press Conference With Elon Musk Friday

May 30, 2025

Andy Ogles Discusses Calling For House Investigations Into Nashville Mayor Critical Of ICE Operations

May 30, 2025

Anti-Israel Protesters Arrested After Disrupting Gal Gadot Film Shoot

May 30, 2025

Subscribe to Updates

Get the latest politics news and updates directly to your inbox.

© 2025 Prices.com LLC. All Rights Reserved.
  • Privacy Policy
  • Terms of use
  • For Advertisers
  • Contact

Type above and press Enter to search. Press Esc to cancel.