Zhengzhong Zhu , Chujun Pu , Xuejie Zhang , Jin Wang , Xiaobing Zhou
{"title":"Dual-dimensional contrastive learning for incomplete multi-view clustering","authors":"Zhengzhong Zhu , Chujun Pu , Xuejie Zhang , Jin Wang , Xiaobing Zhou","doi":"10.1016/j.neucom.2024.128892","DOIUrl":null,"url":null,"abstract":"<div><div>Incomplete multi-view clustering (IMVC) is a critical task in real-world applications, where missing data in some views can severely limit the ability to leverage complementary information across views. This issue leads to incomplete sample representations, hindering model performance. Current contrastive learning methods for IMVC exacerbate the problem by directly constructing data pairs from incomplete samples, ignoring essential information and resulting in class collisions, where samples from different classes are incorrectly grouped together due to a lack of label guidance. These challenges are particularly detrimental in fields like recommendation systems and bioinformatics, where accurate clustering of high-dimensional and incomplete data is essential for decision-making. To address these issues, we propose Dual-dimensional Contrastive Learning (DCL), an online IMVC model that fills missing values through multi-view consistency transfer, enabling simultaneous clustering and representation learning via instance-level and cluster-level contrastive learning in both row and column spaces. DCL mitigates class collision issues by generating high-confidence pseudo-labels and using an optimal transport matrix, significantly improving clustering accuracy. Extensive experiments demonstrate that DCL achieves state-of-the-art results across five datasets. The code is available at <span><span>https://github.com/2251821381/DCL</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"615 ","pages":"Article 128892"},"PeriodicalIF":5.5000,"publicationDate":"2024-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231224016631","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Incomplete multi-view clustering (IMVC) is a critical task in real-world applications, where missing data in some views can severely limit the ability to leverage complementary information across views. This issue leads to incomplete sample representations, hindering model performance. Current contrastive learning methods for IMVC exacerbate the problem by directly constructing data pairs from incomplete samples, ignoring essential information and resulting in class collisions, where samples from different classes are incorrectly grouped together due to a lack of label guidance. These challenges are particularly detrimental in fields like recommendation systems and bioinformatics, where accurate clustering of high-dimensional and incomplete data is essential for decision-making. To address these issues, we propose Dual-dimensional Contrastive Learning (DCL), an online IMVC model that fills missing values through multi-view consistency transfer, enabling simultaneous clustering and representation learning via instance-level and cluster-level contrastive learning in both row and column spaces. DCL mitigates class collision issues by generating high-confidence pseudo-labels and using an optimal transport matrix, significantly improving clustering accuracy. Extensive experiments demonstrate that DCL achieves state-of-the-art results across five datasets. The code is available at https://github.com/2251821381/DCL.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.