Cross-UNet: dual-branch infrared and visible image fusion framework based on cross-convolution and attention mechanism

Published: 01 Jan 2023, Last Modified: 13 Nov 2024Vis. Comput. 2023EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Existing infrared and visible image fusion methods suffer from edge information loss, artifact introduction, and image distortion. Therefore, a dual-branch network model based on the attention mechanism, Cross-UNet, is proposed in this paper for infrared and visible image fusion. First, the encoder part adopts an asymmetric convolution kernel, which can simultaneously obtain local detail information and global structural information of the source image from different directions. Second, in order to fuse the dual-branch image features of different scales, a dual-attention mechanism is added to the fusion block. Finally, the decoder adopts an attention model with a large receptive field to enhance the ability to judge the importance of features, thereby improving the fusion quality. On the public datasets of TNO, RoadScene, and Country, the results are fully compared with nine other advanced fusion methods both qualitatively and quantitatively. The results show that the model in this paper has superior performance and high stability.
Loading