
Modelo de vídeo com IA de próxima geração com narrativa multi-cena, saída 2K e identidade persistente de personagem pela ByteDance.
Seedance 2.0 é o mais recente modelo de geração de vídeo com IA da ByteDance. Ele introduz narrativa multi-cena para sequências de vídeo coerentes, saída de resolução 2K, co-geração nativa de áudio e vídeo e identidade persistente de personagem entre cenas. Comparado ao Seedance 1.5 Pro, oferece renderização 30% mais rápida e controle criativo significativamente aprimorado.
Key specifications of the Seedance 2.0 model.
Max Resolution
Multi-Shot Sequences
Max Duration
Define the narrative arc and plan multiple connected shots for your story.
Upload reference images and configure persistent character identity across shots.
Generate a cinematic multi-shot video with consistent characters and 2K output.
The @ reference system lets you assign specific roles to each uploaded file: @face for character likeness, @motion for movement style, @style for visual tone, @audio for soundtrack sync. No other model offers this level of compositional control over multimodal inputs.
Dolly zooms, rack focuses, tracking shots, POV switches, smooth handheld — Seedance 2.0 scored 9/10 for camera control in benchmark testing, the highest among competing models. Each shot in a multi-shot sequence can have its own camera behavior.
ByteDance incorporated physics-aware training that penalizes impossible motion during generation. Cloth drapes and wrinkles naturally, water splashes with correct weight, collisions have impact, and characters shift balance when walking.
The Dual-Branch Diffusion Transformer generates audio and video in parallel, not sequentially. Spoken dialogue syncs to lip movement at the phoneme level, foley effects match on-screen action, and environmental ambience adapts to scene changes.
Characters retain their facial features, clothing, hairstyle, and proportions across all shots in a multi-shot video. The internal reference-locking system ensures the same person looks the same regardless of camera angle or scene change.
Generate up to 6 independently controlled camera cuts in a single generation. Each shot can have its own framing, camera movement, and action — the output looks like an edited sequence, not a raw single-take generation.
Multi-shot sequences, physics-aware motion, and quad-modal input — all generated by Seedance models with no post-editing or compositing.






Multi-shot 2K video with quad-modal input, persistent character identity, and joint audio generation. 150 credits per 5 seconds.
1080p video generation with audio
4K video generation model
Video generation with audio support
Turbo Pro video generation
AI image generation model
Next-gen AI image generation
4K AI image generation
AI image editing model
Ultra-fast AI image generation