Align then Fusion: Generalized Large-scale Multi-view Clustering with Anchor Matching CorrespondencesDownload PDF

Published: 31 Oct 2022, 18:00, Last Modified: 13 Oct 2022, 03:54NeurIPS 2022 AcceptReaders: Everyone
Keywords: multi-view graph clustering, anchor graph clustering
TL;DR: We present the first tudy of generalized multi-view anchor graph clusterng framework to address the Anchor-Unaligned Problem in large-scale MVC tasks and extensive experiments demonstrate its effectiveness and efficiency.
Abstract: Multi-view anchor graph clustering selects representative anchors to avoid full pair-wise similarities and therefore reduce the complexity of graph methods. Although widely applied in large-scale applications, existing approaches do not pay sufficient attention to establishing correct correspondences between the anchor sets across views. To be specific, anchor graphs obtained from different views are not aligned column-wisely. Such an Anchor-Unaligned Problem (AUP) would cause inaccurate graph fusion and degrade the clustering performance. Under multi-view scenarios, generating correct correspondences could be extremely difficult since anchors are not consistent in feature dimensions. To solve this challenging issue, we propose the first study of the generalized and flexible anchor graph fusion framework termed Fast Multi-View Anchor-Correspondence Clustering (FMVACC). Specifically, we show how to find anchor correspondence with both feature and structure information, after which anchor graph fusion is performed column-wisely. Moreover, we theoretically show the connection between FMVACC and existing multi-view late fusion and partial view-aligned clustering, which further demonstrates our generality. Extensive experiments on seven benchmark datasets demonstrate the effectiveness and efficiency of our proposed method. Moreover, the proposed alignment module also shows significant performance improvement applying to existing multi-view anchor graph competitors indicating the importance of anchor alignment. Our code is available at \url{}.
Supplementary Material: zip
12 Replies