ByteDance’s next-gen AI model can generate clips based on text, images, audio, and video
TikTok's parent company just dropped a multi-modal video bomb that's terrifying Hollywood.
ByteDance has launched Seedance 2.0, a next-gen AI video model that can generate 15-second clips with audio by combining up to nine images, three video clips, and three audio prompts. It uniquely accounts for camera movement, visual effects, and motion to create complex, multi-subject scenes. The model, available on ByteDance's Dreamina platform, is already being used to create hyper-realistic videos featuring celebrities and copyrighted characters, sparking industry concern.
Why It Matters
This multi-modal leap directly challenges OpenAI's Sora and could democratize high-quality video production, disrupting film and content creation.