ChangeDiff: A Multi-Temporal Change Detection Data Generator with Flexible Text Prompts via Diffusion Model
Abstract: progress in change detection (CD) with the support of pixel
level annotations. However, collecting diverse data and man
ually annotating them is costly, laborious, and knowledge
intensive. Existing generative methods for CD data synthe
sis show competitive potential in addressing this issue but
still face the following limitations: 1) difficulty in flexibly
controlling change events, 2) dependence on additional data
to train the data generators, 3) focus on specific change de
tection tasks. To this end, this paper focuses on the seman
tic CD (SCD) task and develops a multi-temporal SCD data
generator ChangeDiff by exploring powerful diffusion mod
els. ChangeDiff innovatively generates change data in two
steps: first, it uses text prompts and a text-to-layout (T2L)
model to create continuous layouts, and then it employs
layout-to-image (L2I) to convert these layouts into images.
Specifically, we propose multi-class distribution-guided text
prompts (MCDG-TP), allowing for layouts to be generated
f
lexibly through controllable classes and their correspond
ing ratios. Subsequently, to generalize the T2L model to the
proposed MCDG-TP, a class distribution refinement loss is
further designed as training supervision. Our generated data
shows significant progress in temporal continuity, spatial di
versity, and quality realism, empowering change detectors
with accuracy and transferability.
Loading