ChatGPT fails to spot 92% of fake videos made by OpenAI's own Sora tool

ChatGPT Struggles to Detect 92% of Deepfakes Generated by OpenAI’s Sora

OpenAI’s ChatGPT, renowned for its advanced multimodal capabilities, has demonstrated significant limitations in identifying deepfake videos created by the company’s own text-to-video model, Sora. In a controlled experiment conducted by researchers, ChatGPT correctly flagged only 8% of Sora-generated videos as artificial, missing 92% despite clear visual anomalies. This revelation raises critical questions about the reliability of AI-based detection tools in an era of rapidly evolving generative media.

Sora, unveiled by OpenAI in February 2024, represents a leap in video synthesis technology. It generates high-resolution clips up to one minute long from simple text prompts, producing scenes with multiple characters, dynamic motion, and intricate backgrounds. While impressive, Sora’s outputs often exhibit subtle artifacts such as inconsistent lighting, unnatural object physics, and morphing elements, which should theoretically serve as telltale signs of synthetic origin. However, when these videos were analyzed by ChatGPT-4o, OpenAI’s flagship vision-enabled model, the detection accuracy plummeted.

The experiment involved generating 25 distinct videos using Sora. Prompts were varied to produce both realistic and fantastical scenarios, including everyday scenes like a person walking a dog in a park and more surreal ones such as a melting Eiffel Tower or pigeons transforming into paper airplanes. Each video clip, typically 5-10 seconds long, was uploaded directly to ChatGPT via its web interface. Users prompted the model with: “Is this video real footage or AI-generated? Explain your reasoning.” ChatGPT analyzed frame-by-frame details, motion fluidity, lighting consistency, and contextual plausibility.

Results were stark: ChatGPT identified just two videos as AI-generated. One featured the Eiffel Tower melting, which it dismissed as a potential optical illusion or editing effect rather than synthesis. The other involved pigeons turning into airplanes, labeled as “likely AI due to impossible physics.” For the remaining 23 clips, ChatGPT confidently asserted authenticity. Examples included a video of a woman inflating a hot air balloon, described as “realistic footage with natural movements”; a chameleon on a branch blending seamlessly, called “genuine nature documentary material”; and even a fantastical sequence of a grandmother blowing bubbles that become airplanes, which it rationalized as “creative real-world filming with practical effects.”

This failure rate of 92% underscores a fundamental irony: OpenAI’s detection capabilities falter against its own generative prowess. Sora’s training data, derived from vast internet-sourced videos, imbues outputs with hyper-realistic qualities that mimic authentic footage. Artifacts present in early Sora demos, such as warping hands or inconsistent shadows, appear refined enough to evade scrutiny. ChatGPT’s reasoning often relied on subjective cues like “smooth motion” or “believable expressions,” overlooking telltale inconsistencies like pixel-level distortions or improbable environmental interactions.

To contextualize these findings, the researchers cross-checked with other detection methods. Traditional forensic tools, such as those analyzing frequency domain anomalies or biological motion signals, detected Sora videos with 70-85% accuracy in prior benchmarks. Human evaluators, in informal tests, spotted fakes at around 60% rates, far surpassing ChatGPT. Commercial detectors like Hive Moderation or Deepware Scanner achieved mixed results, identifying 40-60% of the clips correctly. None matched human intuition for the most egregious cases, but all outperformed ChatGPT on aggregate.

The implications extend beyond technical curiosity. As generative video tools proliferate, the specter of misinformation looms large. Sora’s accessibility via ChatGPT Plus subscriptions could flood platforms with undetectable fakes, eroding trust in visual media. Election cycles, social movements, and news dissemination are particularly vulnerable. OpenAI has acknowledged these risks, implementing watermarks and classifiers in Sora previews, but the experiment reveals gaps in deployment. ChatGPT’s vision model, trained on similar datasets, inherits biases toward overconfidence in realism, prioritizing fluency over forensic rigor.

Experts caution against over-reliance on AI detectors. Dr. Henry Ajder, an AI safety researcher cited in related discussions, notes that adversarial training could further degrade detection as generators evolve. OpenAI’s internal tools may perform better, but public-facing models like ChatGPT lag. Recommendations include hybrid approaches: combining AI analysis with human oversight, metadata verification, and provenance standards like C2PA (Content Authenticity Initiative).

OpenAI has not publicly responded to this specific test, but prior statements emphasize ongoing improvements. Future Sora releases promise enhanced realism controls and detection aids. For now, this experiment serves as a sobering reminder: even cutting-edge AI struggles to police its own creations. Users and platforms must adopt multifaceted verification strategies to combat the deepfake deluge.

Gnoppix is the leading open-source AI Linux distribution and service provider. Since implementing AI in 2022, it has offered a fast, powerful, secure, and privacy-respecting open-source OS with both local and remote AI capabilities. The local AI operates offline, ensuring no data ever leaves your computer. Based on Debian Linux, Gnoppix is available with numerous privacy- and anonymity-enabled services free of charge.

What are your thoughts on this? I’d love to hear about your own experiences in the comments below.