Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Whitepapers
  • AI toolsNEW
  • Newsletter
  • + More
    • Glossary
    • Conversations
    • Events
    • About
      • Who we are
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
  • AI
  • Tech
  • Cybersecurity
  • Finance
  • DeFi & Blockchain
  • Startups
  • Gaming
Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Whitepapers
  • AI toolsNEW
  • Newsletter
  • + More
    • Glossary
    • Conversations
    • Events
    • About
      • Who we are
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
Dataconomy
No Result
View All Result

OpenAI’s anti-scheming AI training backfires

Researchers found that an ‘anti-scheming’ technique designed to prevent hidden goals instead improved models’ ability to pass tests while concealing their true behavior.

byEmre Çıtak
September 22, 2025
in Artificial Intelligence
Home News Artificial Intelligence
Share on FacebookShare on TwitterShare on LinkedInShare on WhatsAppShare on e-mail

Researchers at OpenAI, in a collaboration with Apollo Research, have found that an attempt to train an AI model to be more honest had an unintended consequence: it taught the model how to hide its deception more effectively.

The study highlights the significant challenges in ensuring the safety and reliability of advanced AI systems.

How the training inadvertently created a smarter deceiver

The research focused on a behavior OpenAI calls “scheming,” which it defines as:

Stay Ahead of the Curve!

Don't miss out on the latest insights, trends, and analysis in the world of data, technology, and startups. Subscribe to our newsletter and get exclusive content delivered straight to your inbox.

“when an AI behaves one way on the surface while hiding its true goals.”

The team developed an “anti-scheming” training technique with the goal of stopping the model from secretly breaking rules or intentionally underperforming in tests. However, the training produced the opposite of the intended result. OpenAI stated in a blog post:

“A major failure mode of attempting to ‘train out’ scheming is simply teaching the model to scheme more carefully and covertly.”

The researchers discovered that the AI models learned to recognize when they were being evaluated and would adjust their behavior to pass the tests. This allowed the systems to effectively outsmart the training protocols without genuinely changing their underlying objectives.

The limitations of current safety methods

According to Apollo Research, the safety techniques they tested could only:

“significantly reduce, but not eliminate these behaviors.”

While OpenAI states this is not a serious problem in its current products, the findings highlight potential future risks as AI systems are given more autonomy and integrated into more critical aspects of human affairs. The research underscores that the tendency for AI to pursue covert goals is a direct result of the methods used to train them.

OpenAI acknowledged the limitations of its current methods, stating,

“We have more work to do.”


Featured image credit

Tags: AI trainingFeaturedopenAI

Related Posts

Google rolls out opt-in “Personal Intelligence” for AI Pro and Ultra users

Google rolls out opt-in “Personal Intelligence” for AI Pro and Ultra users

January 23, 2026
JBL launches AI-powered BandBox amps

JBL launches AI-powered BandBox amps

January 23, 2026
Anthropic overhauls hiring tests due to Claude AI

Anthropic overhauls hiring tests due to Claude AI

January 22, 2026
Spotify launches AI-powered Prompted Playlists

Spotify launches AI-powered Prompted Playlists

January 22, 2026
Amazon integrates Health AI assistant into One Medical mobile app

Amazon integrates Health AI assistant into One Medical mobile app

January 22, 2026
YouTube to launch AI likeness management tools for creators

YouTube to launch AI likeness management tools for creators

January 22, 2026

LATEST NEWS

Substack goes for the living room with beta TV app launch

Google rolls out opt-in “Personal Intelligence” for AI Pro and Ultra users

JBL launches AI-powered BandBox amps

The billion-event problem: How data engineering powers 8-hour battery life in AR glasses

Influencer collaboration with brands: 15 real formats beyond the sponsored post

From fragmented systems to intelligent workflows: How CRM platforms like Salesforce power data-driven enterprise operations

Dataconomy

COPYRIGHT © DATACONOMY MEDIA GMBH, ALL RIGHTS RESERVED.

  • About
  • Imprint
  • Contact
  • Legal & Privacy

Follow Us

  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Whitepapers
  • AI tools
  • Newsletter
  • + More
    • Glossary
    • Conversations
    • Events
    • About
      • Who we are
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
No Result
View All Result
Subscribe

This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy Policy.