Graph Contrastive-and-Reconstructive Hashing for Unsupervised Cross-Modal Retrieval

Published: 2025, Last Modified: 15 Jan 2026Data Sci. Eng. 2025EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Hashing-based unsupervised cross-modal retrieval has gained significant attention in the big data management community due to its low storage overhead and rapid retrieval speed. However, current methods often lack effective alignment strategies to reduce the modality gap. They also fail to explore the latent structural information of the training data for accurate relationship learning, resulting in sub-optimal cross-modal retrieval performance. To tackle these challenges, we propose a novel unsupervised cross-modal hashing method called Graph Contrastive-and-Reconstructive Hashing (GCRH). Specifically, GCRH first performs global graph contrastive learning, which involves both intra-modal and inter-modal pairs. This facilitates the learning of more discriminative hash codes through intra-modal discrimination and inter-modal alignment objectives. To further bridge the modality gap, GCRH conducts local graph reconstruction using GCN-based decoders to reconstruct the original features of one modality from the hash codes of another. The integration of contrastive-and-reconstructive learning with graph structural information enables GCRH to generate high-quality hash codes that are both well-aligned and discriminative. Extensive experiments on three benchmark datasets substantiate the superior cross-modal retrieval performance of GCRH.
Loading