Ning Ruan , Yufei Zeng , Huan Liu , Dong Liu , Pengfei Jiao
{"title":"高效异构图少射学习的对比蒸馏","authors":"Ning Ruan , Yufei Zeng , Huan Liu , Dong Liu , Pengfei Jiao","doi":"10.1016/j.neucom.2025.131493","DOIUrl":null,"url":null,"abstract":"<div><div>Heterogeneous graphs (HGs), as a general modeling paradigm for multi-typed entities and complex interactions in ubiquitous real-world networks, have attracted extensive research enthusiasm. While self-supervised learning (SSL) on heterogeneous graph neural networks (HGNNs) demonstrates promising performance, existing SSL approaches face critical limitations: (1) The inherent multiple relation types and meta-path aggregations in HGNNs create prohibitive training and inference costs that restrict scalability; (2) The local message-passing paradigm confines information propagation to immediate neighborhoods, limiting the model’s ability to capture long-range dependencies and global structural patterns essential for complex reasoning tasks; (3) Task specialization necessitates costly fine-tuning of the HGNN backbones. To address these issues, we propose Efficient Heterogeneous Graph Few-shot Learning (EHGFL) to improve HGNNs’ scalability and global-structure modeling capabilities. Specifically, EHGFL first employs instance discrimination contrastive learning for self-supervised pretraining of HGNNs. To enhance efficiency, we introduce a novel cross-model contrastive distillation mechanism that transfers HGNNs’ heterogeneous structure modeling ability to a concise, globally-structure-aware multilayer perceptron. This feature-space distillation process preserves heterogeneous structure representations while avoiding expensive neighborhood aggregation and enhancing global feature awareness. Furthermore, to bridge the gap between pretraining objectives and downstream tasks, we adopt prompt tuning techniques specifically designed for the student model, enabling effective adaptation with limited labeled data. Extensive experiments on two real-world HG datasets demonstrate that the proposed EHGFL framework substantially accelerates training and inference while achieving superior few-shot node classification accuracy compared to state-of-the-art baselines.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"656 ","pages":"Article 131493"},"PeriodicalIF":6.5000,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"EHGFL: Contrastive distillation for efficient heterogeneous graph few-shot learning\",\"authors\":\"Ning Ruan , Yufei Zeng , Huan Liu , Dong Liu , Pengfei Jiao\",\"doi\":\"10.1016/j.neucom.2025.131493\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Heterogeneous graphs (HGs), as a general modeling paradigm for multi-typed entities and complex interactions in ubiquitous real-world networks, have attracted extensive research enthusiasm. While self-supervised learning (SSL) on heterogeneous graph neural networks (HGNNs) demonstrates promising performance, existing SSL approaches face critical limitations: (1) The inherent multiple relation types and meta-path aggregations in HGNNs create prohibitive training and inference costs that restrict scalability; (2) The local message-passing paradigm confines information propagation to immediate neighborhoods, limiting the model’s ability to capture long-range dependencies and global structural patterns essential for complex reasoning tasks; (3) Task specialization necessitates costly fine-tuning of the HGNN backbones. To address these issues, we propose Efficient Heterogeneous Graph Few-shot Learning (EHGFL) to improve HGNNs’ scalability and global-structure modeling capabilities. Specifically, EHGFL first employs instance discrimination contrastive learning for self-supervised pretraining of HGNNs. To enhance efficiency, we introduce a novel cross-model contrastive distillation mechanism that transfers HGNNs’ heterogeneous structure modeling ability to a concise, globally-structure-aware multilayer perceptron. This feature-space distillation process preserves heterogeneous structure representations while avoiding expensive neighborhood aggregation and enhancing global feature awareness. Furthermore, to bridge the gap between pretraining objectives and downstream tasks, we adopt prompt tuning techniques specifically designed for the student model, enabling effective adaptation with limited labeled data. Extensive experiments on two real-world HG datasets demonstrate that the proposed EHGFL framework substantially accelerates training and inference while achieving superior few-shot node classification accuracy compared to state-of-the-art baselines.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"656 \",\"pages\":\"Article 131493\"},\"PeriodicalIF\":6.5000,\"publicationDate\":\"2025-09-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231225021654\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225021654","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
EHGFL: Contrastive distillation for efficient heterogeneous graph few-shot learning
Heterogeneous graphs (HGs), as a general modeling paradigm for multi-typed entities and complex interactions in ubiquitous real-world networks, have attracted extensive research enthusiasm. While self-supervised learning (SSL) on heterogeneous graph neural networks (HGNNs) demonstrates promising performance, existing SSL approaches face critical limitations: (1) The inherent multiple relation types and meta-path aggregations in HGNNs create prohibitive training and inference costs that restrict scalability; (2) The local message-passing paradigm confines information propagation to immediate neighborhoods, limiting the model’s ability to capture long-range dependencies and global structural patterns essential for complex reasoning tasks; (3) Task specialization necessitates costly fine-tuning of the HGNN backbones. To address these issues, we propose Efficient Heterogeneous Graph Few-shot Learning (EHGFL) to improve HGNNs’ scalability and global-structure modeling capabilities. Specifically, EHGFL first employs instance discrimination contrastive learning for self-supervised pretraining of HGNNs. To enhance efficiency, we introduce a novel cross-model contrastive distillation mechanism that transfers HGNNs’ heterogeneous structure modeling ability to a concise, globally-structure-aware multilayer perceptron. This feature-space distillation process preserves heterogeneous structure representations while avoiding expensive neighborhood aggregation and enhancing global feature awareness. Furthermore, to bridge the gap between pretraining objectives and downstream tasks, we adopt prompt tuning techniques specifically designed for the student model, enabling effective adaptation with limited labeled data. Extensive experiments on two real-world HG datasets demonstrate that the proposed EHGFL framework substantially accelerates training and inference while achieving superior few-shot node classification accuracy compared to state-of-the-art baselines.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.