Most creative teams already have the still image. The product shot is done, the portrait approved, the illustration finalized. Seedance v1.0 Lite Image-to-Video accepts that finished image alongside a text prompt and generates a video clip where the subject moves, the camera shifts, and the scene evolves while maintaining visual consistency with the original photograph.
This differs from text-to-video (T2V) generation. A T2V model generates every visual element from scratch based on text alone. With image-to-video (I2V), the source image anchors the output. Colors, textures, spatial layout, and subject appearance carry over from the photograph. The model adds plausible motion and temporal continuity on top of that foundation.
Seedance v1.0 Lite Image-to-Video shares the unified pre-training framework described in the Seedance 1.0 tech report. That architecture handles both T2V and I2V through binary task masks, so the motion modeling, structural stability, and multi-shot awareness that earned Seedance first-place rankings on Artificial Analysis leaderboards are present here. Output spans 720p and 1080p at 24 fps, with durations from two to 12 seconds and seven aspect ratios: 16:9, 9:16, 1:1, 4:3, 3:4, 21:9, and 9:21.
See https://console.byteplus.com/ark/region:ark+ap-southeast-1/model/detail?Id=seedance-1-0-lite-i2v for methodology and benchmark context.
The Lite tier targets volume. Teams animating entire product catalogs or generating dozens of social video variants from one approved photograph can do so at a lower per-clip cost than Pro-tier options.