Seedance 2.0: AI Video Generation with Joint Audio and Video
Coming soonByteDance's Seedance 2.0 brings state-of-the-art text-to-video and image-to-video with optional, high-quality audio — all in one model.
Official Seed2 articleUnlike models that rely only on text or a single image, Seedance 2.0 is multimodal: it accepts images, videos, audio, and text as inputs. You can set the visual style with a reference image, drive motion and camera work with a video, add rhythm with audio, and guide the story with a prompt—giving you control that was previously difficult in generative video. Output is typically 4–15 seconds with native sound effects and music, and the model supports multiple resolutions (480p, 720p, 1080p) with or without audio.
The model is known for physics-aware motion, character and object consistency across frames, and the ability to replicate camera moves and choreography from reference videos. It can extend existing clips, replace characters while keeping the action, and generate audio-synced dialogue and beat-synced edits—useful for ads, localization, and storyboard-to-video workflows.
Seedance 2.0 is coming soon to Hocha. In the meantime, you can use Seedance 1.5 Pro in the AI Video Generator: pick your model, set resolution and options, describe your scene, and generate. Credits are used per second of video, so you only pay for what you create.
Whether you need a quick clip for social media, a concept for a longer piece, or storyboard-style motion, Seedance 2.0 on Hocha gives you a simple way to go from idea to video without leaving the browser.
Ready to create your own AI video?
Try AI Video Generator