Bringing Your Own View: Graph Contrastive Learning without Prefabricated Data Augmentations.

Yuning You, Tianlong Chen, Zhangyang Wang, Yang Shen
{"title":"Bringing Your Own View: Graph Contrastive Learning without Prefabricated Data Augmentations.","authors":"Yuning You, Tianlong Chen, Zhangyang Wang, Yang Shen","doi":"10.1145/3488560.3498416","DOIUrl":null,"url":null,"abstract":"<p><p>Self-supervision is recently surging at its new frontier of graph learning. It facilitates graph representations beneficial to downstream tasks; but its success could hinge on domain knowledge for handcraft or the often expensive trials and errors. Even its state-of-the-art representative, graph contrastive learning (GraphCL), is not completely free of those needs as GraphCL uses a prefabricated prior reflected by the ad-hoc manual selection of graph data augmentations. Our work aims at advancing GraphCL by answering the following questions: <i>How to represent the space of graph augmented views? What principle can be relied upon to learn a prior in that space? And what framework can be constructed to learn the prior in tandem with contrastive learning?</i> Accordingly, we have extended the prefabricated discrete prior in the augmentation set, to a learnable continuous prior in the parameter space of graph generators, assuming that graph priors <i>per se</i>, similar to the concept of image manifolds, can be learned by data generation. Furthermore, to form contrastive views without collapsing to trivial solutions due to the prior learnability, we have leveraged both principles of information minimization (InfoMin) and information bottleneck (InfoBN) to regularize the learned priors. Eventually, contrastive learning, InfoMin, and InfoBN are incorporated organically into one framework of bi-level optimization. Our principled and automated approach has proven to be competitive against the state-of-the-art graph self-supervision methods, including GraphCL, on benchmarks of small graphs; and shown even better generalizability on large-scale graphs, without resorting to human expertise or downstream validation. Our code is publicly released at https://github.com/Shen-Lab/GraphCL_Automated.</p>","PeriodicalId":74530,"journal":{"name":"Proceedings of the ... International Conference on Web Search & Data Mining. International Conference on Web Search & Data Mining","volume":"2022 ","pages":"1300-1309"},"PeriodicalIF":0.0000,"publicationDate":"2022-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9130056/pdf/nihms-1808195.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the ... International Conference on Web Search & Data Mining. International Conference on Web Search & Data Mining","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3488560.3498416","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2022/2/15 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Self-supervision is recently surging at its new frontier of graph learning. It facilitates graph representations beneficial to downstream tasks; but its success could hinge on domain knowledge for handcraft or the often expensive trials and errors. Even its state-of-the-art representative, graph contrastive learning (GraphCL), is not completely free of those needs as GraphCL uses a prefabricated prior reflected by the ad-hoc manual selection of graph data augmentations. Our work aims at advancing GraphCL by answering the following questions: How to represent the space of graph augmented views? What principle can be relied upon to learn a prior in that space? And what framework can be constructed to learn the prior in tandem with contrastive learning? Accordingly, we have extended the prefabricated discrete prior in the augmentation set, to a learnable continuous prior in the parameter space of graph generators, assuming that graph priors per se, similar to the concept of image manifolds, can be learned by data generation. Furthermore, to form contrastive views without collapsing to trivial solutions due to the prior learnability, we have leveraged both principles of information minimization (InfoMin) and information bottleneck (InfoBN) to regularize the learned priors. Eventually, contrastive learning, InfoMin, and InfoBN are incorporated organically into one framework of bi-level optimization. Our principled and automated approach has proven to be competitive against the state-of-the-art graph self-supervision methods, including GraphCL, on benchmarks of small graphs; and shown even better generalizability on large-scale graphs, without resorting to human expertise or downstream validation. Our code is publicly released at https://github.com/Shen-Lab/GraphCL_Automated.

Abstract Image

Abstract Image

自带视图:无需预制数据增强的图形对比学习
自我监督(Self-supervision)最近在图学习的新前沿领域风起云涌。它有助于生成有利于下游任务的图形表征;但其成功与否可能取决于手工领域知识,或者往往是代价高昂的试验和错误。即使是其最先进的代表--图形对比学习(GraphCL),也不能完全摆脱这些需求,因为 GraphCL 使用的是预制先验,反映在图形数据增强的临时手动选择上。我们的工作旨在通过回答以下问题来推动 GraphCL 的发展:如何表示图增强视图的空间?在该空间中学习先验可以依靠什么原则?可以构建怎样的框架来结合对比学习学习先验?因此,我们将增强集中预制的离散先验扩展为图生成器参数空间中可学习的连续先验,假设图先验本身与图像流形的概念类似,可以通过数据生成来学习。此外,为了形成对比视图,同时不因先验的可学习性而坍缩为琐碎的解决方案,我们利用了信息最小化(InfoMin)和信息瓶颈(InfoBN)这两个原理来规范所学的先验。最终,对比学习、InfoMin 和 InfoBN 被有机地整合到一个双层优化框架中。事实证明,在小型图的基准测试中,我们的原则性自动方法与包括 GraphCL 在内的最先进图自监督方法相比具有竞争力;在大型图上,我们的方法显示出更好的通用性,而无需借助人类专业知识或下游验证。我们的代码已在 https://github.com/Shen-Lab/GraphCL_Automated 上公开发布。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信