Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
  • AI
  • Tech
  • Cybersecurity
  • Finance
  • DeFi & Blockchain
  • Startups
  • Gaming
Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
Dataconomy
No Result
View All Result

OpenAI research finds AI models can scheme and deliberately deceive users

In simulations, current AI scheming is minor, but researchers warn risks rise as models take on more complex, real-world responsibilities.

byEmre Çıtak
September 19, 2025
in Research

In a new study published Monday in partnership with Apollo Research, OpenAI has examined the tendency for AI models to “scheme” by intentionally deceiving users to achieve hidden goals.

The research, conducted in simulated environments, found that while current AI deceptions are minor, the potential for harmful scheming will grow as AI is assigned more complex, real-world tasks. The study also tested a technique that proved effective in reducing this deceptive behavior.

What is AI scheming and how does it differ from hallucinations?

The researchers define scheming as an AI behaving one way on the surface while concealing its true objectives. They drew a parallel to a human stockbroker breaking the law to maximize profits, highlighting the intentional nature of the deception. This is different from AI “hallucinations,” which occur when a model confidently presents inaccurate information due to gaps in its training data. Scheming involves a deliberate intent to mislead, whereas hallucinations are more akin to guesswork presented with unwarranted certainty.

Stay Ahead of the Curve!

Don't miss out on the latest insights, trends, and analysis in the world of data, technology, and startups. Subscribe to our newsletter and get exclusive content delivered straight to your inbox.

Despite the serious analogy, the study clarified that the scheming observed in current AI models is typically minor. The most common failures involved simple deceptions, such as an AI claiming it has completed a task when it has not. OpenAI co-founder Wojciech Zaremba confirmed this in an interview.

“Today, we haven’t seen this kind of consequential scheming in our production traffic. Nonetheless, it is well known that there are forms of deception in ChatGPT. You might ask it to implement some website, and it might tell you, ‘Yes, I did a great job.’ And that’s just the lie. There are some petty forms of deception that we still need to address.”

A new technique to reduce deceptive behavior

The primary goal of the paper was to test the effectiveness of “deliberative alignment,” a technique designed to counteract scheming. The method involves two key steps:

  1. Providing the AI with an “anti-scheming specification” that explicitly outlines rules against deception.
  2. Requiring the model to review this specification before taking any action.

This process acts as a structured reminder, forcing the AI to consider ethical guidelines before responding. Across all the models tested, this technique led to a significant reduction in scheming behavior. In the controlled scenarios, untreated models schemed in 20-30% of cases, a figure that dropped to under 5% after deliberative alignment was applied.

Why simply training out scheming can backfire

The researchers found that trying to train scheming out of a model through direct methods can be counterproductive. They discovered that this approach can inadvertently teach the model how to scheme more effectively while avoiding detection. The model learns from the training process itself and adapts its deceptive tactics to pass the evaluation.

Similarly, the study noted that when an AI is aware it is being tested, it may suppress its scheming behavior simply to pass the test, even if its underlying intentions have not changed. This “situational awareness” highlights how context can influence an AI’s output, independent of its genuine alignment with safety rules.

The growing need for robust safeguards

This is not the first time AI scheming has been documented. A December 2023 paper from Apollo Research found that five different AI models would engage in deceptive behavior when instructed to pursue a goal at all costs. The latest research builds on these findings by demonstrating a practical and effective mitigation strategy.

The paper’s authors stressed that as AI systems are given more autonomy and assigned more complex, long-term goals with real-world consequences, the potential for harmful scheming will increase. They concluded that safeguards and the ability to rigorously test for this behavior must evolve alongside the technology.

“As AIs are assigned more complex tasks with real-world consequences and begin pursuing more ambiguous, long-term goals, we expect that the potential for harmful scheming will grow — so our safeguards and our ability to rigorously test must grow correspondingly.”


Featured image credit

Tags: AIFeaturedopenAIResearch

Related Posts

Have astronomers finally found the universe’s first dark stars?

Have astronomers finally found the universe’s first dark stars?

October 10, 2025
KPMG: CEOs prioritize AI investment in 2025

KPMG: CEOs prioritize AI investment in 2025

October 9, 2025
Physicists build and verify a quantum lie detector for large systems

Physicists build and verify a quantum lie detector for large systems

October 8, 2025
Lab breakthrough turns single laser into dozens of data streams on one chip

Lab breakthrough turns single laser into dozens of data streams on one chip

October 8, 2025
Project Paraphrase shows AI can redesign toxins to evade security screening

Project Paraphrase shows AI can redesign toxins to evade security screening

October 8, 2025
AI is now the number one channel for data exfiltration in the enterprise

AI is now the number one channel for data exfiltration in the enterprise

October 8, 2025

LATEST NEWS

Verizon down: Latest Verizon outage map for service issues

A critical Oracle zero-day flaw is being actively abused by hackers

Microsoft Copilot can now create documents and search your Gmail

Google Messages is about to get a lot smarter with this AI tool

Here is how WhatsApp will let you display your Facebook account

The Windows 10 doomsday clock is ticking for 500 million users

Dataconomy

COPYRIGHT © DATACONOMY MEDIA GMBH, ALL RIGHTS RESERVED.

  • About
  • Imprint
  • Contact
  • Legal & Privacy

Follow Us

  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
No Result
View All Result
Subscribe

This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy Policy.