REEDIT: Multimodal Exemplar-Based Image Editing

Published: 01 Jan 2025, Last Modified: 25 Jun 2025WACV 2025EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Modern Text-to-Image (T2I) Diffusion models have revolutionized image editing by enabling the generation of high-quality photorealistic images. While the de-facto method for performing edits with T2I models is through text instructions, this approach is non-trivial due to the complex many-to-many mapping between natural language and images. In this work, we address exemplar-based image editing - the task of transferring an edit from an exemplar pair to a content image(s). We propose Reedit, a modular and efficient end-to-end framework that captures edits in both text and image modalities while ensuring the fidelity of the edited image. We validate the effectiveness of Reedit through extensive comparisons with state-of-the-art baselines and sensitivity analyses of key design choices. Our results demonstrate that Reeditconsistently outperforms contemporary approaches both qualitatively and quantitatively. Additionally, Reedit boasts high practical applicability, as it does not require any task-specific optimization and is 4× faster than the existing state-of-the-art. The code and data for our work is available at https://reedit-diffusion.github.io/.
Loading