Abstract
Benefiting from the strong view-consistent information mining capacity, multi-view contrastive clustering has attracted plenty of attention in recent years. However, we observe the following drawback, which limits the clustering performance from further improvement. The existing multi-view models mainly focus on the consistency of the same samples in different views while ignoring the circumstance of similar but different samples in cross-view scenarios. To solve this problem, we propose a novel Dual contrastive calibration network for Multi-View Clustering (DealMVC). Specifically, we first design a fusion mechanism to obtain a global cross-view feature. Then, a global contrastive calibration loss is proposed by aligning the view feature similarity graph and the high-confidence pseudo-label graph. Moreover, to utilize the diversity of multi-view information, we propose a local contrastive calibration loss to constrain the consistency of pair-wise view features. The feature structure is regularized by reliable class information, thus guaranteeing similar samples have similar features in different views. During the training procedure, the interacted cross-view feature is jointly optimized at both local and global levels. In comparison with other state-of-the-art approaches, the comprehensive experimental results obtained from eight benchmark datasets provide substantial validation of the effectiveness and superiority of our algorithm. We release the code of DealMVC at https://github.com/xihongyang1999/DealMVC on GitHub.
Author supplied keywords
Cite
CITATION STYLE
Yang, X., Jiaqi, J., Wang, S., Liang, K., Liu, Y., Wen, Y., … Zhu, E. (2023). DealMVC: Dual Contrastive Calibration for Multi-view Clustering. In MM 2023 - Proceedings of the 31st ACM International Conference on Multimedia (pp. 337–346). Association for Computing Machinery, Inc. https://doi.org/10.1145/3581783.3611951
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.