高效异构图少射学习的对比蒸馏

IF 6.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Ning Ruan , Yufei Zeng , Huan Liu , Dong Liu , Pengfei Jiao
{"title":"高效异构图少射学习的对比蒸馏","authors":"Ning Ruan ,&nbsp;Yufei Zeng ,&nbsp;Huan Liu ,&nbsp;Dong Liu ,&nbsp;Pengfei Jiao","doi":"10.1016/j.neucom.2025.131493","DOIUrl":null,"url":null,"abstract":"<div><div>Heterogeneous graphs (HGs), as a general modeling paradigm for multi-typed entities and complex interactions in ubiquitous real-world networks, have attracted extensive research enthusiasm. While self-supervised learning (SSL) on heterogeneous graph neural networks (HGNNs) demonstrates promising performance, existing SSL approaches face critical limitations: (1) The inherent multiple relation types and meta-path aggregations in HGNNs create prohibitive training and inference costs that restrict scalability; (2) The local message-passing paradigm confines information propagation to immediate neighborhoods, limiting the model’s ability to capture long-range dependencies and global structural patterns essential for complex reasoning tasks; (3) Task specialization necessitates costly fine-tuning of the HGNN backbones. To address these issues, we propose Efficient Heterogeneous Graph Few-shot Learning (EHGFL) to improve HGNNs’ scalability and global-structure modeling capabilities. Specifically, EHGFL first employs instance discrimination contrastive learning for self-supervised pretraining of HGNNs. To enhance efficiency, we introduce a novel cross-model contrastive distillation mechanism that transfers HGNNs’ heterogeneous structure modeling ability to a concise, globally-structure-aware multilayer perceptron. This feature-space distillation process preserves heterogeneous structure representations while avoiding expensive neighborhood aggregation and enhancing global feature awareness. Furthermore, to bridge the gap between pretraining objectives and downstream tasks, we adopt prompt tuning techniques specifically designed for the student model, enabling effective adaptation with limited labeled data. Extensive experiments on two real-world HG datasets demonstrate that the proposed EHGFL framework substantially accelerates training and inference while achieving superior few-shot node classification accuracy compared to state-of-the-art baselines.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"656 ","pages":"Article 131493"},"PeriodicalIF":6.5000,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"EHGFL: Contrastive distillation for efficient heterogeneous graph few-shot learning\",\"authors\":\"Ning Ruan ,&nbsp;Yufei Zeng ,&nbsp;Huan Liu ,&nbsp;Dong Liu ,&nbsp;Pengfei Jiao\",\"doi\":\"10.1016/j.neucom.2025.131493\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Heterogeneous graphs (HGs), as a general modeling paradigm for multi-typed entities and complex interactions in ubiquitous real-world networks, have attracted extensive research enthusiasm. While self-supervised learning (SSL) on heterogeneous graph neural networks (HGNNs) demonstrates promising performance, existing SSL approaches face critical limitations: (1) The inherent multiple relation types and meta-path aggregations in HGNNs create prohibitive training and inference costs that restrict scalability; (2) The local message-passing paradigm confines information propagation to immediate neighborhoods, limiting the model’s ability to capture long-range dependencies and global structural patterns essential for complex reasoning tasks; (3) Task specialization necessitates costly fine-tuning of the HGNN backbones. To address these issues, we propose Efficient Heterogeneous Graph Few-shot Learning (EHGFL) to improve HGNNs’ scalability and global-structure modeling capabilities. Specifically, EHGFL first employs instance discrimination contrastive learning for self-supervised pretraining of HGNNs. To enhance efficiency, we introduce a novel cross-model contrastive distillation mechanism that transfers HGNNs’ heterogeneous structure modeling ability to a concise, globally-structure-aware multilayer perceptron. This feature-space distillation process preserves heterogeneous structure representations while avoiding expensive neighborhood aggregation and enhancing global feature awareness. Furthermore, to bridge the gap between pretraining objectives and downstream tasks, we adopt prompt tuning techniques specifically designed for the student model, enabling effective adaptation with limited labeled data. Extensive experiments on two real-world HG datasets demonstrate that the proposed EHGFL framework substantially accelerates training and inference while achieving superior few-shot node classification accuracy compared to state-of-the-art baselines.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"656 \",\"pages\":\"Article 131493\"},\"PeriodicalIF\":6.5000,\"publicationDate\":\"2025-09-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231225021654\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225021654","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

异构图(HGs)作为普遍存在的现实世界网络中多类型实体和复杂交互的通用建模范式,引起了广泛的研究热情。虽然异构图神经网络(hgnn)上的自监督学习(SSL)表现出了良好的性能,但现有的SSL方法面临着严重的局限性:(1)hgnn中固有的多关系类型和元路径聚合产生了过高的训练和推理成本,限制了可扩展性;(2)本地消息传递范式将信息传播限制在邻近区域,限制了模型捕获复杂推理任务所必需的远程依赖关系和全局结构模式的能力;(3)任务专门化需要对HGNN主干进行昂贵的微调。为了解决这些问题,我们提出了高效异构图少射学习(EHGFL)来提高hgnn的可扩展性和全局结构建模能力。具体来说,EHGFL首先采用实例辨别对比学习对hgnn进行自监督预训练。为了提高效率,我们引入了一种新的跨模型对比蒸馏机制,将hgnn的异质结构建模能力转移到一个简洁的、全局结构感知的多层感知器上。这种特征空间蒸馏过程保留了异构结构表示,同时避免了昂贵的邻域聚合,增强了全局特征感知。此外,为了弥合预训练目标和下游任务之间的差距,我们采用了专门为学生模型设计的快速调优技术,能够有效地适应有限的标记数据。在两个真实HG数据集上进行的大量实验表明,与最先进的基线相比,所提出的EHGFL框架大大加速了训练和推理,同时实现了优越的少射节点分类精度。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
EHGFL: Contrastive distillation for efficient heterogeneous graph few-shot learning
Heterogeneous graphs (HGs), as a general modeling paradigm for multi-typed entities and complex interactions in ubiquitous real-world networks, have attracted extensive research enthusiasm. While self-supervised learning (SSL) on heterogeneous graph neural networks (HGNNs) demonstrates promising performance, existing SSL approaches face critical limitations: (1) The inherent multiple relation types and meta-path aggregations in HGNNs create prohibitive training and inference costs that restrict scalability; (2) The local message-passing paradigm confines information propagation to immediate neighborhoods, limiting the model’s ability to capture long-range dependencies and global structural patterns essential for complex reasoning tasks; (3) Task specialization necessitates costly fine-tuning of the HGNN backbones. To address these issues, we propose Efficient Heterogeneous Graph Few-shot Learning (EHGFL) to improve HGNNs’ scalability and global-structure modeling capabilities. Specifically, EHGFL first employs instance discrimination contrastive learning for self-supervised pretraining of HGNNs. To enhance efficiency, we introduce a novel cross-model contrastive distillation mechanism that transfers HGNNs’ heterogeneous structure modeling ability to a concise, globally-structure-aware multilayer perceptron. This feature-space distillation process preserves heterogeneous structure representations while avoiding expensive neighborhood aggregation and enhancing global feature awareness. Furthermore, to bridge the gap between pretraining objectives and downstream tasks, we adopt prompt tuning techniques specifically designed for the student model, enabling effective adaptation with limited labeled data. Extensive experiments on two real-world HG datasets demonstrate that the proposed EHGFL framework substantially accelerates training and inference while achieving superior few-shot node classification accuracy compared to state-of-the-art baselines.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Neurocomputing
Neurocomputing 工程技术-计算机:人工智能
CiteScore
13.10
自引率
10.00%
发文量
1382
审稿时长
70 days
期刊介绍: Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信