Abstract: Predicting structured outputs such as semantic segmentation relies on expensive per-pixel annotations to learn strong supervised models like convolutional neural networks. However, these models trained on one data domain may not generalize well to other domains unequipped with annotations for model finetuning. To avoid the labor-intensive process of annotation, we develop a domain adaptation method to adapt the source data to the unlabeled target domain. To this end, we propose to learn discriminative feature representations of patches based on label histograms in the source domain, through the construction of a disentangled space. With such representations as guidance, we then use an adversarial learning scheme to push the feature representations in target patches to the closer distributions in source ones. In addition, we show that our framework can integrate a global alignment process with the proposed patch-level alignment and achieve state-of-the-art performance on semantic segmentation. Extensive ablation studies and experiments are conducted on numerous benchmark datasets with various settings, such as synthetic-to-real and cross-city scenarios.
Keywords: Domain Adaptation, Feature Representation Learning, Semantic Segmentation
TL;DR: A domain adaptation method for structured output via learning patch-level discriminative feature representations
Data: [Cityscapes](https://paperswithcode.com/dataset/cityscapes), [GTA5](https://paperswithcode.com/dataset/gta5), [SYNTHIA](https://paperswithcode.com/dataset/synthia)
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 8 code implementations](https://www.catalyzex.com/paper/domain-adaptation-for-structured-output-via/code)
14 Replies
Loading