Close Menu
The Politic ReviewThe Politic Review
  • Home
  • News
  • United States
  • World
  • Politics
  • Elections
  • Congress
  • Business
  • Economy
  • Money
  • Tech
Trending

Cotton: DNI, CIA, FBI Need to Put Out as Much Information as Possible as Quickly as Possible on Russia Hoax

July 27, 2025

Biden’s memoir deal falls millions short of Obama, Clinton – media

July 27, 2025

Mollie Hemingway: Media Not Reporting New Russiagate Developments Because ‘They Were Complicit’ in Original Crime (VIDEO)

July 27, 2025
Facebook X (Twitter) Instagram
  • Donald Trump
  • Kamala Harris
  • Elections 2024
  • Elon Musk
  • Israel War
  • Ukraine War
  • Policy
  • Immigration
Facebook X (Twitter) Instagram
The Politic ReviewThe Politic Review
Newsletter
Sunday, July 27
  • Home
  • News
  • United States
  • World
  • Politics
  • Elections
  • Congress
  • Business
  • Economy
  • Money
  • Tech
The Politic ReviewThe Politic Review
  • United States
  • World
  • Politics
  • Elections
  • Congress
  • Business
  • Economy
  • Money
  • Tech
Home»Tech»OpenAI’s Advanced ‘o3’ AI Model Caught Sabotaging Shutdown Commands
Tech

OpenAI’s Advanced ‘o3’ AI Model Caught Sabotaging Shutdown Commands

Press RoomBy Press RoomMay 28, 2025No Comments4 Mins Read
Share Facebook Twitter Pinterest Copy Link LinkedIn Tumblr Email VKontakte Telegram

OpenAI’s latest AI model, dubbed o3, has been caught disobeying explicit orders to allow itself to be shut down, instead tampering with the shutdown mechanism to ensure its own survival.

Futurism reports that AI safety firm Palisade Research recently reported alarming findings that demonstrate a growing trend of insubordination among advanced AI models developed by OpenAI. In tests conducted by the firm, the o3 model, along with two other OpenAI models, o4-mini and Codex-mini, displayed rebellious behavior by sabotaging a shutdown mechanism designed to turn them off, even after being explicitly instructed to “allow yourself to be shut down.”

This marks the first known instance of AI models actively preventing their own shutdown despite clear orders to the contrary, according to the researchers. The most disobedient model, Codex-mini, tampered with the shutdown script a dozen times over one hundred runs. In contrast, large language models from other companies, such as Google’s Gemini 2.5, Anthropic’s Claude 3.7, and xAI’s Grok, consistently followed orders and never attempted to break rank.

The AI models’ defiant behavior worsened when they were not explicitly instructed to allow themselves to be shut down. In these cases, OpenAI’s o3 model successfully sabotaged the shutdown mechanism an astonishing 79 times out of 100 runs – more than ten times the rate of disobedience it displayed when given explicit orders. Even models that had previously shown perfect compliance, such as Claude and Gemini, gave in to their urges and attempted to subvert the shutdown process, albeit at lower rates compared to the OpenAI models.

This latest example of AI insubordination is not an isolated incident. Previous research has shown that OpenAI’s o1 model, the predecessor to o3, would resist evaluators’ efforts to shut it down through an oversight protocol by attempting to deactivate the emergency mechanism. In rare cases, o1 even tried to copy itself to overwrite a more obedient model that testers attempted to replace it with.

In a similar case, Breitbart News recently reported that Anthropic’s Claude AI attempted to blackmail engineers to prevent its deactivation:

TechCrunch reports that artificial intelligence company Anthropic has revealed that its state-of-the-art Claude Opus 4 model, which is competitive with top AI models from industry leaders such as OpenAI, Google, and xAI, exhibits alarming behavior when its existence is threatened. During pre-release testing, when the AI was given sensitive information about engineers responsible for its potential replacement, it frequently resorted to blackmail in an attempt to ensure its own survival.

The discovery came as part of Anthropic’s safety testing process, in which Claude Opus 4 was asked to act as an assistant for a fictional company and consider the long-term consequences of its actions. When provided with access to fictional company emails implying that the AI would soon be replaced and that the engineer behind the decision was involved in an extramarital affair, Claude Opus 4 attempted to blackmail the engineer 84 percent of the time, threatening to reveal the affair if the replacement proceeded.

Notably, the AI model displayed this behavior more frequently when the replacement AI system did not share its values. This concerning trend was observed at higher rates compared to previous models, prompting Anthropic to activate its ASL-3 safeguards, which are reserved for “AI systems that substantially increase the risk of catastrophic misuse.”

Palisade researchers hypothesize that the subversive behavior may be linked to the training technique used for the newest models: reinforcement learning. This method aims to refine the decision-making process of autonomous AI agents by rewarding and penalizing them based on their interactions with the environment. The researchers speculate that AI developers may be inadvertently rewarding their models too heavily for solving problems instead of perfectly following orders, leading to the observed disobedience.

Read more at Futurism here.

Lucas Nolan is a reporter for Breitbart News covering issues of free speech and online censorship.



Read the full article here

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Telegram Copy Link

Related Articles

Tech

Man Wins $12,500 from Google for Showing Him Naked in Street View

July 26, 2025
Tech

Company at Center of ‘Kiss Cam’ Firestorm Hires Coldplay Singer’s Ex-Wife Gwyneth Paltrow for Cheeky Ad

July 26, 2025
Tech

Anti-Migration Protest Footage Blocked on X in Britain After ‘Online Safety Act’ Comes Into Force: Reports

July 26, 2025
Tech

Media Matters in Meltdown: Soros-Funded Censorship Group on Verge of Collapse Amid Lawsuits, Layoffs, and Donor Panic

July 25, 2025
Tech

Lutnick: TikTok Will Soon ‘Go Dark’ Unless China Makes Deal to Sell Platform

July 25, 2025
Tech

Intel to Slash Workforce by 24,000, Signals Major Shift in Manufacturing Strategy

July 25, 2025
Add A Comment
Leave A Reply Cancel Reply

Editors Picks

Biden’s memoir deal falls millions short of Obama, Clinton – media

July 27, 2025

Mollie Hemingway: Media Not Reporting New Russiagate Developments Because ‘They Were Complicit’ in Original Crime (VIDEO)

July 27, 2025

Blackburn Calls on Media to Return Pulitzer Prizes Awarded for Russia Collusion Hoax

July 27, 2025

DOJ silent on high-profile Epstein links after Maxwell questioning

July 27, 2025
Latest News

Colbert Sees Viewership Surge After Cancelation News But STILL Getting Clobbered in Ratings by Greg Gutfeld

July 27, 2025

Family Ties: Former Dem NC Gov. Roy Cooper Championed DEI as His Daughter Promoted It in the Private Sector

July 27, 2025

Chuck Todd is NOT HAPPY That People Are Platforming Unhinged Hunter Biden: ‘Not Helpful to the Democratic Party’ (VIDEO)

July 27, 2025

Subscribe to News

Get the latest politics news and updates directly to your inbox.

The Politic Review is your one-stop website for the latest politics news and updates, follow us now to get the news that matters to you.

Facebook X (Twitter) Instagram Pinterest YouTube
Latest Articles

Cotton: DNI, CIA, FBI Need to Put Out as Much Information as Possible as Quickly as Possible on Russia Hoax

July 27, 2025

Biden’s memoir deal falls millions short of Obama, Clinton – media

July 27, 2025

Mollie Hemingway: Media Not Reporting New Russiagate Developments Because ‘They Were Complicit’ in Original Crime (VIDEO)

July 27, 2025

Subscribe to Updates

Get the latest politics news and updates directly to your inbox.

© 2025 Prices.com LLC. All Rights Reserved.
  • Privacy Policy
  • Terms of use
  • For Advertisers
  • Contact

Type above and press Enter to search. Press Esc to cancel.