Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
  • AI
  • Tech
  • Cybersecurity
  • Finance
  • DeFi & Blockchain
  • Startups
  • Gaming
Dataconomy
  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
Subscribe
No Result
View All Result
Dataconomy
No Result
View All Result

OpenAI faces criticism after CTO’s interview on Sora

OpenAI CTO Mira Murati's statements about SORA's training data, or rather ones that she could not make, drew massive reaction

byEmre Çıtak
March 15, 2024
in Artificial Intelligence
Home News Artificial Intelligence

OpenAI, the influential artificial intelligence research lab behind groundbreaking tools like ChatGPT and Sora, has found itself in hot water following a recent interview with its Chief Technology Officer, Mira Murati.

The interview, conducted by Wall Street Journal reporter Joanna Stern, focused on OpenAI’s latest image, or rather video, generation system, Sora.

Concerns center around the potential misuse of copyrighted work to train AI models and the lack of transparency from OpenAI regarding its data practices.

Stay Ahead of the Curve!

Don't miss out on the latest insights, trends, and analysis in the world of data, technology, and startups. Subscribe to our newsletter and get exclusive content delivered straight to your inbox.

Sora’s training data is in question

At the heart of the controversy lies the issue of training data, the massive datasets used to train AI models.

When asked about the sources of data utilized for Sora, Murati provided the standard response: the model had been trained on “publicly available and licensed data“.

However, further probing revealed hesitation and uncertainty on Murati’s part about the specific details of this dataset.

This response has raised red flags among artists, photographers, and intellectual property experts. AI image generation systems depend heavily on ingesting vast quantities of images, many of which may be protected by copyright. The lack of clarity around Sora’s training data raises questions about whether OpenAI has adequately safeguarded the rights of content creators.

OpenAI SORA training data controversy
Sora’s training database has not been published on any official platform (Image credit)

Shutterstock usage admitted later on

Adding fuel to the fire was Murati’s initial refusal to address whether Shutterstock images were a component of Sora’s training dataset. Only after the interview, in a footnote added by the Wall Street Journal, did Murati confirm the use of Shutterstock’s image library.

This confirmation contradicts OpenAI’s public-facing stance of “publicly available and licensed data” and suggests an attempt to conceal potentially problematic sourcing practices.

Shutterstock and OpenAI formed a partnership granting OpenAI rights to use Shutterstock’s image library in training image generation models like DALL-E 2 and potentially Sora.

In return, Shutterstock contributors (the photographers and artists whose images are on the platform) receive compensation when their work is used in the development of these AI models.

A PR nightmare unfolds

It’s safe to say that most public relations folks would not consider this interview to be a PR masterpiece.

Murati’s lack of clarity comes at a sensitive time for OpenAI, already facing major copyright lawsuits, including a significant one filed by the New York Times.

The public is scrutinizing practices like OpenAI’s alleged secret use of YouTube videos for model training, as previously reported by The Information. With stakeholders ranging from artists to politicians demanding accountability, Murati’s avoidance only fuels the fire.

OpenAI’s opaque approach is backfiring spectacularly, transforming the Sora interview into a PR disaster.

OpenAI CTO Mira Murati says Sora was trained on publicly available and licensed data pic.twitter.com/rf7pZ0ZX00

— Tsarathustra (@tsarnick) March 13, 2024

Transparency is not the most discussed topic for nothing

This incident underscores a critical truth: unveiling the truth is paramount in the world of AI. OpenAI’s stumbling responses have severely undermined public trust and intensified questions about its ethical practices. The Sora controversy highlights the growing chorus demanding greater accountability within the AI industry.

Murati’s reluctance to disclose the specifics of Sora’s training data breeds mistrust and sets a dangerous precedent.

Without the clarity artists, creators, and the public are demanding, ethical debates and the potential for legal action will only intensify.

There are no angels in this land

While much of the current scrutiny falls squarely on OpenAI, it’s crucial to remember they’re not the only player in the game.

Facebook AI Research’s LLaMA model and Google’s Gemini have also faced allegations of problematic training data sources.

OpenAI SORA training data controversy
The issue of transparency in AI developments has been on the agenda for a long time (Image credit)

This isn’t surprising, as Business Insider reports that Meta has already admitted to using Instagram and Facebook posts to train its AI models. Additionally, Google’s control over vast swaths of the internet gives them unparalleled access to potential training data, raising similar ethical concerns about consent and copyright.

The situation with OpenAI’s Sora is just one piece of a larger puzzle. The entire AI development field is facing scrutiny regarding its data practices and the potential ethical implications.


Featured image credit: Freepik.

Tags: DatabaseFeaturedopenAIsora

Related Posts

AI chatbots spread false info in 1 of 3 responses

AI chatbots spread false info in 1 of 3 responses

September 5, 2025
OpenAI to mass produce custom AI chip with Broadcom in 2025

OpenAI to mass produce custom AI chip with Broadcom in 2025

September 5, 2025
Deepmind finds RAG limit with fixed-size embeddings

Deepmind finds RAG limit with fixed-size embeddings

September 5, 2025
TCL QM9K integrates Gemini with presence detection

TCL QM9K integrates Gemini with presence detection

September 5, 2025
LunaLock ransomware hits artists/clients with AI training threat

LunaLock ransomware hits artists/clients with AI training threat

September 5, 2025
OpenAI: New ‘OpenAI for Science’ uses GPT-5

OpenAI: New ‘OpenAI for Science’ uses GPT-5

September 5, 2025

LATEST NEWS

Texas Attorney General files lawsuit over the PowerSchool data breach

iPhone 17 Pro is expected to arrive with 48mp telephoto, variable aperture expected

AI chatbots spread false info in 1 of 3 responses

OpenAI to mass produce custom AI chip with Broadcom in 2025

When two Mark Zuckerbergs collide

Deepmind finds RAG limit with fixed-size embeddings

Dataconomy

COPYRIGHT © DATACONOMY MEDIA GMBH, ALL RIGHTS RESERVED.

  • About
  • Imprint
  • Contact
  • Legal & Privacy

Follow Us

  • News
    • Artificial Intelligence
    • Cybersecurity
    • DeFi & Blockchain
    • Finance
    • Gaming
    • Startups
    • Tech
  • Industry
  • Research
  • Resources
    • Articles
    • Guides
    • Case Studies
    • Glossary
    • Whitepapers
  • Newsletter
  • + More
    • Conversations
    • Events
    • About
      • About
      • Contact
      • Imprint
      • Legal & Privacy
      • Partner With Us
No Result
View All Result
Subscribe

This website uses cookies. By continuing to use this website you are giving consent to cookies being used. Visit our Privacy Policy.