Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Whitepapers
  • AI toolsNEW
  • Newsletter
  • + More
    • Glossary
    • Conversations
    • Events
    • About
      • Who we are
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
  • AI
  • Tech
  • Cybersecurity
  • Finance
  • DeFi & Blockchain
  • Startups
  • Gaming
Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Whitepapers
  • AI toolsNEW
  • Newsletter
  • + More
    • Glossary
    • Conversations
    • Events
    • About
      • Who we are
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
Dataconomy
No Result
View All Result

Elon Musk’s xAI chatbot Grok exposed hundreds of thousands of private user conversations

Exposed content included instructions for illegal activities, malware code, assassination plots, passwords, and medical inquiries.

byKerem Gülen
September 19, 2025
in Cybersecurity, Artificial Intelligence
Home News Cybersecurity
Share on FacebookShare on TwitterShare on LinkedInShare on WhatsAppShare on e-mail

Elon Musk’s xAI chatbot, Grok, has inadvertently exposed hundreds of thousands of private user conversations to public search engines due to a flaw in its “share” function. The incident raises significant privacy and security concerns for the platform, which, according to Musk, has reached 64 million monthly users. This user base is smaller than competitors like OpenAI’s ChatGPT (700 million weekly users) and Google’s Gemini (450 million users).

How the Grok privacy breach occurred

The exposure happened through Grok’s built-in feature for sharing conversations, which creates a unique URL for each shared dialogue. These URLs were subsequently indexed by search engines like Google, making the private conversations publicly searchable online. Users were not explicitly warned that their shared chats could become part of a public database.

What information was exposed?

The publicly indexed conversations contained a wide range of sensitive and dangerous information, including:

Stay Ahead of the Curve!

Don't miss out on the latest insights, trends, and analysis in the world of data, technology, and startups. Subscribe to our newsletter and get exclusive content delivered straight to your inbox.

  • Detailed instructions for illegal activities, such as how to manufacture fentanyl and explosives.
  • Code for writing malware.
  • Assassination plots, including one that specifically targeted Elon Musk.
  • Personal information, such as passwords and detailed medical inquiries about health conditions and treatments.

While some of the illicit prompts may have been from security researchers testing the chatbot’s limits, the sheer volume of exposed private data has triggered alarm over the platform’s design and its privacy safeguards.

Grok’s performance and ongoing content issues

Grok’s latest version, Grok 4, has shown improved performance on some benchmarks and occasionally uses Elon Musk’s social media posts to generate answers on controversial topics. However, the model has also produced problematic content, including antisemitic remarks. These issues have created challenges for xAI as it seeks to integrate Grok more deeply into Musk’s other companies, such as Tesla and the social media platform X.

xAI positions Grok as a premium chatbot, charging users a hefty $300 monthly subscription for SuperGrok and offering an API for enterprise clients.

However, persistent concerns about its content alignment and erratic behavior remain potential obstacles to its broader adoption. The company has not yet issued a public statement addressing the exposed conversations or the privacy breach.


Featured image credit

Tags: grokxai

Related Posts

Anthropic partners with Teach For All to train 100,000 global educators

Anthropic partners with Teach For All to train 100,000 global educators

January 20, 2026
Signal co-founder launches privacy-focused AI service Confer

Signal co-founder launches privacy-focused AI service Confer

January 20, 2026
Adobe launches AI-powered Object Mask for Premiere Pro

Adobe launches AI-powered Object Mask for Premiere Pro

January 20, 2026
Claim: NVIDIA green-lit pirated book downloads for AI training

Claim: NVIDIA green-lit pirated book downloads for AI training

January 20, 2026
OpenAI targets “practical adoption” for 2026 strategy

OpenAI targets “practical adoption” for 2026 strategy

January 20, 2026
Mother of one of Elon Musk’s children sues xAI over sexual Grok deepfakes

Mother of one of Elon Musk’s children sues xAI over sexual Grok deepfakes

January 16, 2026

LATEST NEWS

Anthropic partners with Teach For All to train 100,000 global educators

Signal co-founder launches privacy-focused AI service Confer

Adobe launches AI-powered Object Mask for Premiere Pro

Google Workspace adds password-protected Office file editing

Claim: NVIDIA green-lit pirated book downloads for AI training

Tesla restarts Dojo3 supercomputer project as AI5 chip stabilizes

Dataconomy

COPYRIGHT © DATACONOMY MEDIA GMBH, ALL RIGHTS RESERVED.

  • About
  • Imprint
  • Contact
  • Legal & Privacy

Follow Us

  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Whitepapers
  • AI tools
  • Newsletter
  • + More
    • Glossary
    • Conversations
    • Events
    • About
      • Who we are
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
No Result
View All Result
Subscribe

This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy Policy.