Xin Zhang , Wanyu Chen , Fei Cai , Jianming Zheng , Zhiqiang Pan , Yupu Guo , Honghui Chen
{"title":"图上的信息瓶颈驱动提示,用于统一下游的少量分类任务","authors":"Xin Zhang , Wanyu Chen , Fei Cai , Jianming Zheng , Zhiqiang Pan , Yupu Guo , Honghui Chen","doi":"10.1016/j.ipm.2025.104092","DOIUrl":null,"url":null,"abstract":"<div><div>Inspired by the success of prompt in natural language processing, the graph prompt-based methods are proposed to solve the classification tasks under the conditions with limited instances. Recent graph prompt-based methods typically employ link prediction as the pre-training task, which brings a gap between the pre-training task and the downstream classification task, thus introducing irrelevant and noisy features to the downstream tasks. To tackle this issue, we propose a framework called Diffused Graph Prompt (Di-Graph), which consists of three major components, <em>i.e</em>., the diffused pre-training, a graph prompt layer, and an information bottleneck optimizer. Specifically, the diffused pre-training aims to obtain the stable node features with a diffusion process, mitigating the gap between the pre-training tasks and the downstream tasks. The graph prompt layer enhances the pre-trained model to leverage its knowledge via capturing both the structural and node features in graphs. The information bottleneck optimizer helps the model discard redundant features by retaining the minimal sufficient statistic of the input data. Extensive experimental results on five public graph datasets demonstrate that our Di-Graph model surpasses the state-of-the-art model in terms of accuracy for both node and graph classification tasks. In particular, for graph-level tasks, Di-Graph achieves an average performance gain of 6.50% over the previous best model (GraphCL) on BZR dataset. For node-level tasks, Di-Graph achieves a 2.05% improvement over the best baseline (GraphCL) on PROTEINS dataset.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"62 4","pages":"Article 104092"},"PeriodicalIF":7.4000,"publicationDate":"2025-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Information bottleneck-driven prompt on graphs for unifying downstream few-shot classification tasks\",\"authors\":\"Xin Zhang , Wanyu Chen , Fei Cai , Jianming Zheng , Zhiqiang Pan , Yupu Guo , Honghui Chen\",\"doi\":\"10.1016/j.ipm.2025.104092\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Inspired by the success of prompt in natural language processing, the graph prompt-based methods are proposed to solve the classification tasks under the conditions with limited instances. Recent graph prompt-based methods typically employ link prediction as the pre-training task, which brings a gap between the pre-training task and the downstream classification task, thus introducing irrelevant and noisy features to the downstream tasks. To tackle this issue, we propose a framework called Diffused Graph Prompt (Di-Graph), which consists of three major components, <em>i.e</em>., the diffused pre-training, a graph prompt layer, and an information bottleneck optimizer. Specifically, the diffused pre-training aims to obtain the stable node features with a diffusion process, mitigating the gap between the pre-training tasks and the downstream tasks. The graph prompt layer enhances the pre-trained model to leverage its knowledge via capturing both the structural and node features in graphs. The information bottleneck optimizer helps the model discard redundant features by retaining the minimal sufficient statistic of the input data. Extensive experimental results on five public graph datasets demonstrate that our Di-Graph model surpasses the state-of-the-art model in terms of accuracy for both node and graph classification tasks. In particular, for graph-level tasks, Di-Graph achieves an average performance gain of 6.50% over the previous best model (GraphCL) on BZR dataset. For node-level tasks, Di-Graph achieves a 2.05% improvement over the best baseline (GraphCL) on PROTEINS dataset.</div></div>\",\"PeriodicalId\":50365,\"journal\":{\"name\":\"Information Processing & Management\",\"volume\":\"62 4\",\"pages\":\"Article 104092\"},\"PeriodicalIF\":7.4000,\"publicationDate\":\"2025-02-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Processing & Management\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0306457325000342\",\"RegionNum\":1,\"RegionCategory\":\"管理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Processing & Management","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306457325000342","RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Information bottleneck-driven prompt on graphs for unifying downstream few-shot classification tasks
Inspired by the success of prompt in natural language processing, the graph prompt-based methods are proposed to solve the classification tasks under the conditions with limited instances. Recent graph prompt-based methods typically employ link prediction as the pre-training task, which brings a gap between the pre-training task and the downstream classification task, thus introducing irrelevant and noisy features to the downstream tasks. To tackle this issue, we propose a framework called Diffused Graph Prompt (Di-Graph), which consists of three major components, i.e., the diffused pre-training, a graph prompt layer, and an information bottleneck optimizer. Specifically, the diffused pre-training aims to obtain the stable node features with a diffusion process, mitigating the gap between the pre-training tasks and the downstream tasks. The graph prompt layer enhances the pre-trained model to leverage its knowledge via capturing both the structural and node features in graphs. The information bottleneck optimizer helps the model discard redundant features by retaining the minimal sufficient statistic of the input data. Extensive experimental results on five public graph datasets demonstrate that our Di-Graph model surpasses the state-of-the-art model in terms of accuracy for both node and graph classification tasks. In particular, for graph-level tasks, Di-Graph achieves an average performance gain of 6.50% over the previous best model (GraphCL) on BZR dataset. For node-level tasks, Di-Graph achieves a 2.05% improvement over the best baseline (GraphCL) on PROTEINS dataset.
期刊介绍:
Information Processing and Management is dedicated to publishing cutting-edge original research at the convergence of computing and information science. Our scope encompasses theory, methods, and applications across various domains, including advertising, business, health, information science, information technology marketing, and social computing.
We aim to cater to the interests of both primary researchers and practitioners by offering an effective platform for the timely dissemination of advanced and topical issues in this interdisciplinary field. The journal places particular emphasis on original research articles, research survey articles, research method articles, and articles addressing critical applications of research. Join us in advancing knowledge and innovation at the intersection of computing and information science.