S²ENet: Spatial-Spectral Cross-Modal Enhancement Network for Classification of Hyperspectral and LiDAR Data

Published: 01 Jan 2022, Last Modified: 13 Nov 2024IEEE Geosci. Remote. Sens. Lett. 2022EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: The effective utilization of multimodal data (e.g., hyperspectral and light detection and ranging (LiDAR) data) has profound implications for further development of the remote sensing (RS) field. Many studies have explored how to effectively fuse features from multiple modalities; however, few of them focus on information interactions that can effectively promote the complementary semantic content of multisource data before fusion. In this letter, we propose a spatial–spectral enhancement module (S 2 EM) for cross-modal information interaction in deep neural networks. Specifically, S 2 EM consists of SpAtial Enhancement Module (SAEM) for enhancing spatial representation of hyperspectral data by LiDAR features and SpEctral Enhancement Module (SEEM) for enhancing spectral representation of LiDAR data by hyperspectral features. A series of experiments and ablation studies on the Houston2013 dataset show that S 2 EM can effectively facilitate the interaction and understanding between multimodal data. Our source code is available at https://github.com/likyoo/Multimodal-Remote-Sensing-Toolkit , contributing to the RS community.
Loading