ByteDance announced SeedEdit — a new model for image editing via text prompts. It enables high-quality modifications, including local replacements, geometric transformations, and stylistic adjustments.
SeedEdit is built on a diffusion model that combines image generation with precise editing. ByteDance notes that the new editing approach provides “stable high-quality image editing even without prior preparation.” The technology supports sequential image revisions while preserving their aesthetic appeal.
The main challenge in development was the lack of paired data for training. The team solved this by integrating text-image models as a base editing layer. Then, the model was adapted for precise execution of editing instructions.
SeedEdit demonstrates the ability to handle tasks of any complexity — from style changes to adding fantastical elements. Examples include object replacement, lighting adjustments, and landscape transformations.