MS-Former: Multi-Scale Self-Guided Transformer for Medical Image SegmentationDownload PDF

Published: 04 Apr 2023, Last Modified: 30 Apr 2023MIDL 2023 OralReaders: Everyone
Keywords: Transformer, Inter-scale, Intra-scale, Segmentation, Medical Image
TL;DR: our approach considers the semantic dependency that exists between different scales to generate a supervisory signal for inter-scale consistency and also imposes a spatial stability loss within the scale for self-supervised content clustering.
Abstract: Multi-scale representations have proven to be a powerful tool since they can take into account both the fine-grained details of objects in an image as well as the broader context. Inspired by this, we propose a novel dual-branch transformer network that operates on two different scales to encode global contextual dependencies while preserving local information. To learn in a self-supervised fashion, our approach considers the semantic dependency that exists between different scales to generate a supervisory signal for inter-scale consistency and also imposes a spatial stability loss within the scale for self-supervised content clustering. While intra-scale and inter-scale consistency losses aim to increase features similarly within the cluster, we propose to include a cross-entropy loss function on top of the clustering score map to effectively model each cluster distribution and increase the decision boundary between clusters. Iteratively our algorithm learns to assign each pixel to a semantically related cluster to produce the segmentation map. Extensive experiments on skin lesion and lung segmentation datasets show the superiority of our method compared to the state-of-the-art (SOTA) approaches. The implementation code is publicly available at: https://github.com/mindflow-institue/MS-Former.
4 Replies

Loading