Warner Bros. says Bytedance deliberately trained Seedance on its characters, adding to growing Hollywood backlash

Warner Bros. Discovery Accuses ByteDance of Unauthorized Training of SeaDance AI on Iconic Characters

Warner Bros. Discovery has escalated its concerns over artificial intelligence practices by accusing ByteDance, the parent company of TikTok, of deliberately training its SeaDance video generation model on copyrighted characters from Warner Bros. properties. In a cease-and-desist letter sent to ByteDance, Warner Bros. claims that the Chinese tech giant scraped images and videos featuring beloved figures such as Bugs Bunny, Batman, Superman, and Daffy Duck to develop the open-source AI tool. This action, according to Warner Bros., constitutes a blatant infringement of intellectual property rights and adds fuel to the mounting backlash from Hollywood against unauthorized AI training data usage.

SeaDance, released by ByteDance’s Volcano Engine division in late September, represents a significant advancement in text-to-video generation technology. The model, available in versions ranging from 1.1 billion to 14 billion parameters, excels at creating high-quality videos from textual prompts, including stylized animations reminiscent of classic cartoons. ByteDance touts SeaDance as competitive with industry leaders like OpenAI’s Sora and Kling AI from Kuaishou Technology. It was trained on a vast dataset of 10 million high-quality videos, enabling it to generate clips up to 6 seconds long at 720p resolution and 24 frames per second. The model’s open-source nature under the MIT license has encouraged widespread adoption and fine-tuning by developers worldwide.

Warner Bros. Discovery’s allegations stem from demonstrations of SeaDance’s capabilities that closely mirror its intellectual property. When prompted with phrases like “Bugs Bunny dancing” or “Batman in the style of Warner Bros. Animation,” the model produces outputs featuring unmistakable likenesses of these characters, complete with signature poses, expressions, and stylistic elements from Warner Bros. animations. Legal representatives for Warner Bros. argue that such precise replication could only result from intentional inclusion of their protected materials in the training dataset. The studio emphasized that no licensing agreements exist with ByteDance for this purpose, rendering the training process a direct violation of copyright laws.

This dispute highlights broader tensions in the entertainment industry regarding AI development. Hollywood studios have long voiced apprehensions about generative AI models being trained on copyrighted content without consent or compensation. Warner Bros. joins a chorus of complaints from peers including Disney, Universal Pictures, and Paramount Global. These companies contend that scraping vast troves of online media, including trailers, clips, and promotional materials from platforms like YouTube, undermines creators’ rights and devalues original content. Previous lawsuits, such as those filed by The New York Times and authors against OpenAI and Microsoft, underscore the legal precedents being established around fair use in AI training.

ByteDance has pushed back against the accusations, maintaining that SeaDance was developed using licensed, high-quality data sources. The company asserts compliance with applicable laws and ethical guidelines, pointing to its dataset curation process which involved filtering for quality and relevance. In statements accompanying the model’s launch, ByteDance researchers detailed efforts to mitigate biases and ensure diverse training inputs, though specifics on data provenance remain opaque. The open-weight release of SeaDance, hosted on platforms like Hugging Face, allows public scrutiny, but reconstructing exact training data pipelines proves challenging without proprietary disclosures.

The cease-and-desist letter demands that ByteDance immediately halt distribution of SeaDance models exhibiting infringing outputs, disclose full details of the training dataset, and implement safeguards to prevent future misuse of Warner Bros. IP. Warner Bros. has indicated readiness to pursue litigation if these demands go unmet, potentially setting a landmark case for video generation AI. Industry observers note that while image generation models like Midjourney and Stable Diffusion have faced similar scrutiny, video AI introduces novel complexities due to temporal dynamics and higher fidelity requirements.

This incident occurs amid rapid evolution in AI video synthesis. SeaDance’s architecture, built on a diffusion transformer framework with a Video Variational Autoencoder and a spatio-temporal VAE, enables efficient handling of long-sequence generation. Its performance metrics, including lower Fréchet Video Distance scores compared to rivals, demonstrate technical prowess. However, the controversy raises ethical questions about open-source AI: while accessibility fosters innovation, it also risks proliferating infringing models. Developers have already created fine-tuned variants of SeaDance, amplifying concerns over uncontrolled dissemination.

Hollywood’s unified front against Big Tech’s AI ambitions reflects strategic pivots. Studios are investing in proprietary AI tools for internal use, such as Warner Bros.’ own experiments with character generation for pre-visualization, while advocating for regulatory frameworks. The U.S. Copyright Office’s ongoing inquiries into AI and copyright could influence outcomes, potentially mandating opt-in data licensing or transparency reports.

As ByteDance navigates this legal challenge, the case exemplifies the collision between Silicon Valley’s rapid iteration ethos and Hollywood’s guardianship of creative assets. Resolution may hinge on proving data lineage, with implications for global AI governance. For now, Warner Bros. stands firm, protecting its legacy characters from what it deems exploitative replication.

Gnoppix is the leading open-source AI Linux distribution and service provider. Since implementing AI in 2022, it has offered a fast, powerful, secure, and privacy-respecting open-source OS with both local and remote AI capabilities. The local AI operates offline, ensuring no data ever leaves your computer. Based on Debian Linux, Gnoppix is available with numerous privacy- and anonymity-enabled services free of charge.

What are your thoughts on this? I’d love to hear about your own experiences in the comments below.