{"title":"Community Embeddings with Bayesian Gaussian Mixture Model and Variational Inference","authors":"Anton Begehr, Peter B. Panfilov","doi":"10.1109/CBI54897.2022.10053","DOIUrl":null,"url":null,"abstract":"Graphs, such as social networks, emerge naturally from various real-world situations. Recently, graph embedding methods have gained traction in data science research. The graph and community embedding algorithm ComE aims to preserve first-, second- and higher-order proximity. ComE requires prior knowledge of the number of communities K. In this paper, ComE is extended to utilize a Bayesian Gaussian mixture model with variational inference for learning community embeddings (ComE BGMM+VI), similar to ComE+. ComE BGMM+VI takes K as the maximum number of communities and drops components through the trade-off hyperparameter weight concentration prior. The advantage of ComE BGMM+VI over the non-Bayesian ComE for an unknown number of communities K is shown for the small Karate club dataset and explored for the larger DBLP dataset.","PeriodicalId":447040,"journal":{"name":"2022 IEEE 24th Conference on Business Informatics (CBI)","volume":"215 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 24th Conference on Business Informatics (CBI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CBI54897.2022.10053","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Graphs, such as social networks, emerge naturally from various real-world situations. Recently, graph embedding methods have gained traction in data science research. The graph and community embedding algorithm ComE aims to preserve first-, second- and higher-order proximity. ComE requires prior knowledge of the number of communities K. In this paper, ComE is extended to utilize a Bayesian Gaussian mixture model with variational inference for learning community embeddings (ComE BGMM+VI), similar to ComE+. ComE BGMM+VI takes K as the maximum number of communities and drops components through the trade-off hyperparameter weight concentration prior. The advantage of ComE BGMM+VI over the non-Bayesian ComE for an unknown number of communities K is shown for the small Karate club dataset and explored for the larger DBLP dataset.
图,如社交网络,自然地从各种现实世界的情况中出现。最近,图嵌入方法在数据科学研究中获得了关注。图和社区嵌入算法ComE旨在保持一阶,二阶和高阶接近。ComE需要预先知道社区k的数量。本文将ComE扩展为使用带有变分推理的贝叶斯高斯混合模型来学习社区嵌入(ComE BGMM+VI),类似于ComE+。ComE BGMM+VI以K为最大群落数,通过权衡超参数权重集中先验来降低分量。对于未知数量的社区K, ComE BGMM+VI相对于非贝叶斯ComE的优势在小型空手道俱乐部数据集中得到了体现,并在大型DBLP数据集中得到了探索。