Exploring the Role of Depth Information From DAM in Diverse Remote Sensing Semantic Segmentation Tasks
Abstract: Existing semantic segmentation methods for remote sensing images focus mainly on planar features to boost performance but inadequately consider the potential advantages of incorporating depth features. To address this issue, we integrate monocular depth estimations generated by the depth anything model (DAM) and propose three methods that progressively deepen the utilization of depth information in fully supervised, semisupervised, and unsupervised domain adaptation tasks. Specifically, for fully supervised tasks, depth-aware edge consistency loss is applied to enhance boundary positioning and mitigate the adverse effects of intraclass depth variations; for semisupervised tasks, depth-guided weight update, guided by mixture of experts, injects geometric cues and abstract semantics from depth into unlabeled data; and for unsupervised domain adaptation tasks, a depth-enhanced adversarial feature learning strategy facilitates stereo feature alignment between source and target domains to improve domain adaptation effectiveness. Extensive experiments on various remote sensing datasets demonstrate the effectiveness and generality of the proposed methods, which require no modifications to the model architecture and can be readily adapted to similar tasks, yielding average mIoU gains of 0.93%, 1.47%, and 1.29% on fully supervised, semisupervised, and unsupervised domain adaptation tasks, respectively.
External IDs:doi:10.1109/jstars.2025.3620999
Loading