MDP: A Generalized Framework for Text-Guided Image Editing by Manipulating the Diffusion Path

TMLR Paper2666 Authors

10 May 2024 (modified: 29 Jul 2024)Under review for TMLREveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Image generation using diffusion can be controlled in multiple ways. In this paper, we systematically analyze the equations of modern generative diffusion networks to propose a framework, called MDP, that explains the design space of suitable manipulations. We identify 5 different manipulations, including intermediate latent, conditional embedding, cross attention maps, guidance, and predicted noise. We analyze the corresponding parameters of these manipulations and the manipulation schedule. We show that some previous editing methods fit nicely into our framework. Particularly, we identified one specific configuration as a new type of control by manipulating the predicted noise, which can perform higher-quality edits than previous work for a variety of local and global edits.
Submission Length: Regular submission (no more than 12 pages of main content)
Changes Since Last Submission: - Added additional editing results using Stable Diffusion 2.1. - Added visual comparison between DiffEdit and MDP-$\epsilon_t$. - Added a broader impact section. - Added a discussion of the availability of the initial condition - Added details of user study. - Expanded the discussion of the quantitative results. - Fixed typos and formatting issues.
Assigned Action Editor: ~Shiyu_Chang2
Submission Number: 2666
Loading