RAFT: Robust Augmentation of FeaTures for Image Segmentation

Published: 06 May 2025, Last Modified: 06 May 2025SynData4CVEveryoneRevisionsBibTeXCC BY 4.0
Keywords: syn2real, domain adaptation, active learning, augmentation, image segmentation, augmentation
TL;DR: We propose a novel framework for active domain adaptation from synthetic to real data of image segmentation models.
Abstract: Image segmentation is a powerful computer vision technique for scene understanding. However, real-world deployment is stymied by the need for high-quality, meticulously labeled datasets. Synthetic data provides high-quality labels while reducing the need for manual data collection and annotation. However, deep neural networks trained on synthetic data often face the Syn2Real problem, leading to poor performance in real-world deployments. To mitigate the aforementioned gap in image segmentation, we propose RAFT, a novel framework for adapting image segmentation models using minimal labeled real-world data through data and feature augmentations, as well as active learning. To validate RAFT, we perform experiments on the synthetic-to-real "SYNTHIA$\rightarrow$Cityscapes" and "GTAV$\rightarrow$Cityscapes" benchmarks. We manage to surpass the previous state of the art, HALO. SYNTHIA$\rightarrow$Cityscapes experiences an improvement in mIoU* upon domain adaptation of 2.1\%/79.9\%, and GTAV$\rightarrow$Cityscapes experiences a 0.4\%/78.2\% improvement in mIoU. Furthermore, we test our approach on the real-to-real benchmark of "Cityscapes$\rightarrow$ACDC", and again surpass HALO, with a gain in mIoU upon adaptation of 1.3\%/73.2\%. Finally, we examine the effect of the allocated annotation budget and various components of RAFT upon the final transfer mIoU.
Submission Number: 62
Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview