Multi-modality boundary-guided network for generalizable image manipulation localization

Published: 2025, Last Modified: 04 Nov 2025Multim. Syst. 2025EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: The main concern in image manipulation localization is the development of a feature representation that can effectively detect various manipulation techniques such as copy-move, removal, and splicing. However, many existing techniques focus on identifying specific tampering techniques, which limits their applicability in real-world scenarios. To address this challenge, we propose a new approach called the Multi-Modality Boundary-Guided Network (M2BG-Net) to enhance the generalization capacity of the model. Our approach includes several key components. Firstly, we introduce a multi-modality feature learning module that combines high-frequency features with RGB features. This allows us to detect subtle tampering traces that may not be visible in the RGB domain alone. Additionally, we propose an edge-aware module (EAM) to enhance the boundary features of the tampered region. By focusing on the edges, we can improve the accuracy of tampering localization. Furthermore, we introduce a multi-dilation context aggregation module (MCAM) that jointly optimizes RGB streams, frequency streams, and boundary artifacts. This enables us to improve the manipulation localization capability on unknown forgery patterns. Through extensive experiments, we demonstrate that our proposed M2BG-Net outperforms existing image tampering localization methods in terms of generalization and robustness to various tampering means. Our approach represents a significant step towards more effective and reliable image manipulation detection in real-world scenarios.
Loading