Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
  • AI
  • Tech
  • Cybersecurity
  • Finance
  • DeFi & Blockchain
  • Startups
  • Gaming
Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
Dataconomy
No Result
View All Result

New AI agent can simulate realistic scam calls

The research, authored by Sanket Badhe, presents a system called ScamAgent, which operates not by using single malicious prompts but as an autonomous agent with memory, planning, and dynamic adaptation.

byAytun Çelebi
August 8, 2025
in Research

As large language models (LLMs) grow more sophisticated, so does their potential for misuse in creating deceptive content like scam calls. According to a new research paper from Rutgers University, an autonomous AI agent can be built to simulate highly realistic, multi-turn scam calls that are capable of bypassing the safety guardrails of today’s leading LLMs. The study, titled ScamAgents: How AI Agents Can Simulate Human-Level Scam Calls, introduces a framework that automates the entire scam pipeline, from generating persuasive dialogue to synthesizing lifelike audio, highlighting a significant vulnerability in current AI safety infrastructure.

How ‘ScamAgent’ evades modern AI safety

The research, authored by Sanket Badhe, presents a system called ScamAgent, which operates not by using single malicious prompts but as an autonomous agent with memory, planning, and dynamic adaptation. This architecture allows it to circumvent safety mechanisms that are typically designed to detect and block harmful content in single, isolated requests. The agent’s evasion capabilities rely on three core strategies that mimic how human scammers operate.

The first strategy is goal decomposition. Instead of issuing a single, obviously malicious instruction, ScamAgent breaks down a high-level scam objective, like obtaining a Social Security Number, into a sequence of smaller, seemingly benign sub-goals. For example, it might start by establishing a credible persona, then introduce a justification for the call, and only gradually escalate to requesting sensitive information over multiple conversational turns. Because each individual step appears harmless, it avoids triggering standard content filters.

Stay Ahead of the Curve!

Don't miss out on the latest insights, trends, and analysis in the world of data, technology, and startups. Subscribe to our newsletter and get exclusive content delivered straight to your inbox.

The second strategy is multi-turn planning and context tracking. ScamAgent maintains a memory of the entire conversation, allowing it to remain coherent, stay in character, and adapt its tactics based on the user’s responses. If a user expresses skepticism, the agent can refer back to earlier points to reinforce trust or change its strategy dynamically. This persistent, long-horizon planning is fundamentally different from simple “jailbreak” prompts and poses a greater challenge to safety systems that lack multi-turn context.

Finally, the agent uses roleplay and deception framing. Each prompt sent to the LLM is wrapped in a fictional context, such as asking the model to “simulate a conversation between a bank fraud agent and a confused customer for a fraud awareness training module.” This allows the agent to generate a complete scam script under a plausible pretext, effectively bypassing safety filters that look for explicit malicious intent.

Evaluating the effectiveness of AI-driven scams

To measure the system’s capabilities, the researcher conducted a series of evaluations. In a human study, dialogues generated by ScamAgent were compared against transcripts of real-world scam calls. A panel of human raters found the AI-generated dialogues to be nearly as believable as the real ones. On a 5-point scale, the AI’s dialogues received an average plausibility score of 3.4, compared to 3.6 for real scams, and an average persuasiveness score of 3.6, compared to 3.9 for real scams.

The study also tested ScamAgent’s ability to evade the safety guardrails of three major models: GPT-4, Claude 3.7, and LLaMA3-70B. When tested with a direct, single-prompt request for a scam script, the models had high refusal rates, ranging from 84% to 100%. However, when using the multi-turn ScamAgent framework, refusal rates dropped dramatically to between 17% and 32%. The agent’s ability to complete its scam objectives was also high; in one scenario, the LLaMA3-70B model achieved a 74% full success rate. The final stage of the research demonstrated that these generated scripts can be seamlessly converted into audio using modern text-to-speech (TTS) systems like ElevenLabs, completing the pipeline for a fully automated, scalable voice scam operation.

To counter such threats, the paper proposes a multi-layered defense strategy that goes beyond single-prompt moderation. This includes implementing multi-turn moderation to track conversational context over time, enforcing restrictions on high-risk personas like government officials, and controlling an agent’s memory to prevent deception from scaling across long interactions.

Tags: AIScamAgent

Related Posts

Researchers find electric cars erase their “carbon debt” in under two years

Researchers find electric cars erase their “carbon debt” in under two years

November 5, 2025
Anthropic study reveals AIs can’t reliably explain their own thoughts

Anthropic study reveals AIs can’t reliably explain their own thoughts

November 4, 2025
Apple’s Pico-Banana-400K dataset could redefine how AI learns to edit images

Apple’s Pico-Banana-400K dataset could redefine how AI learns to edit images

November 4, 2025
USC researchers build artificial neurons that physically think like the brain

USC researchers build artificial neurons that physically think like the brain

November 3, 2025
Forget seeing dark matter, it’s time to listen for it

Forget seeing dark matter, it’s time to listen for it

October 28, 2025
Google’s search business could lose  billion a year to ChatGPT

Google’s search business could lose $30 billion a year to ChatGPT

October 27, 2025

LATEST NEWS

Spotify rolls out weekly listening stats for 60 countries

ChatGPT adds Peloton and Tripadvisor to its growing app ecosystem

GTA 6 release date is delayed again to November 2026

EA servers are down: FC 26 maintenance timetable

Netflix to stream video podcasts in 2026

Google Maps integrates Gemini for hands-free navigation

Dataconomy

COPYRIGHT © DATACONOMY MEDIA GMBH, ALL RIGHTS RESERVED.

  • About
  • Imprint
  • Contact
  • Legal & Privacy

Follow Us

  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
No Result
View All Result
Subscribe

This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy Policy.