Influence-Based Mini-Batching for Graph Neural Networks

LOG IN Pub Date : 2022-12-18 DOI:10.48550/arXiv.2212.09083
J. Gasteiger, Chen Qian, Stephan Gunnemann
{"title":"Influence-Based Mini-Batching for Graph Neural Networks","authors":"J. Gasteiger, Chen Qian, Stephan Gunnemann","doi":"10.48550/arXiv.2212.09083","DOIUrl":null,"url":null,"abstract":"Using graph neural networks for large graphs is challenging since there is no clear way of constructing mini-batches. To solve this, previous methods have relied on sampling or graph clustering. While these approaches often lead to good training convergence, they introduce significant overhead due to expensive random data accesses and perform poorly during inference. In this work we instead focus on model behavior during inference. We theoretically model batch construction via maximizing the influence score of nodes on the outputs. This formulation leads to optimal approximation of the output when we do not have knowledge of the trained model. We call the resulting method influence-based mini-batching (IBMB). IBMB accelerates inference by up to 130x compared to previous methods that reach similar accuracy. Remarkably, with adaptive optimization and the right training schedule IBMB can also substantially accelerate training, thanks to precomputed batches and consecutive memory accesses. This results in up to 18x faster training per epoch and up to 17x faster convergence per runtime compared to previous methods.","PeriodicalId":379381,"journal":{"name":"LOG IN","volume":"81 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"LOG IN","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48550/arXiv.2212.09083","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5

Abstract

Using graph neural networks for large graphs is challenging since there is no clear way of constructing mini-batches. To solve this, previous methods have relied on sampling or graph clustering. While these approaches often lead to good training convergence, they introduce significant overhead due to expensive random data accesses and perform poorly during inference. In this work we instead focus on model behavior during inference. We theoretically model batch construction via maximizing the influence score of nodes on the outputs. This formulation leads to optimal approximation of the output when we do not have knowledge of the trained model. We call the resulting method influence-based mini-batching (IBMB). IBMB accelerates inference by up to 130x compared to previous methods that reach similar accuracy. Remarkably, with adaptive optimization and the right training schedule IBMB can also substantially accelerate training, thanks to precomputed batches and consecutive memory accesses. This results in up to 18x faster training per epoch and up to 17x faster convergence per runtime compared to previous methods.
基于影响的图神经网络小批处理
使用图神经网络处理大型图是具有挑战性的,因为没有明确的方法来构建小批量。为了解决这个问题,以前的方法依赖于采样或图聚类。虽然这些方法通常会导致良好的训练收敛,但由于昂贵的随机数据访问,它们带来了巨大的开销,并且在推理期间表现不佳。在这项工作中,我们转而关注推理过程中的模型行为。从理论上讲,我们通过最大化节点对输出的影响评分来建模批量构建。当我们不知道训练模型的知识时,这个公式导致输出的最优逼近。我们称之为基于影响的小批处理(IBMB)。与以前达到类似精度的方法相比,IBMB将推理速度提高了130倍。值得注意的是,通过自适应优化和正确的训练计划,IBMB还可以大大加快训练速度,这要归功于预先计算的批处理和连续的内存访问。与以前的方法相比,每个epoch的训练速度提高了18倍,每次运行时的收敛速度提高了17倍。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信