{"title":"基于潜在空间模型生成链接的大规模信息网络嵌入评价","authors":"Shotaro Kawasaki, Ryosuke Motegi, Shogo Matsuno, Yoichi Seki","doi":"10.1145/3520084.3520111","DOIUrl":null,"url":null,"abstract":"Graph representation learning encodes vertices as low-dimensional vectors that summarize their graph position and the structure of their local graph neighborhood. These methods give us beneficial representation in continuous space from big relational data. However, the algorithms are usually evaluated indirectly from the accuracy of applying the learning results to classification tasks because of not giving the correct answer when graph representation learning is applied. Therefore, this study proposes a method to evaluate graph representation learning algorithms by preparing correct learning results for the data by distributing objects in the latent space in advance and probabilistically generating relational graph data from the distributions in the latent space. Using this method, we evaluated LINE: Large-scale information network embedding, one of the most popular algorithms for learning graph representations. LINE consists of two algorithms optimizing two objective functions defined by first-order proximity and second-order proximity. We prepared two link-generating models suitable for these two objective functions and clarified that the corresponding LINE algorithm performed well for the link data generated by each model.","PeriodicalId":444957,"journal":{"name":"Proceedings of the 2022 5th International Conference on Software Engineering and Information Management","volume":"82 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Evaluation of Large-scale Information Network Embedding based on Latent Space Model Generating Links\",\"authors\":\"Shotaro Kawasaki, Ryosuke Motegi, Shogo Matsuno, Yoichi Seki\",\"doi\":\"10.1145/3520084.3520111\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Graph representation learning encodes vertices as low-dimensional vectors that summarize their graph position and the structure of their local graph neighborhood. These methods give us beneficial representation in continuous space from big relational data. However, the algorithms are usually evaluated indirectly from the accuracy of applying the learning results to classification tasks because of not giving the correct answer when graph representation learning is applied. Therefore, this study proposes a method to evaluate graph representation learning algorithms by preparing correct learning results for the data by distributing objects in the latent space in advance and probabilistically generating relational graph data from the distributions in the latent space. Using this method, we evaluated LINE: Large-scale information network embedding, one of the most popular algorithms for learning graph representations. LINE consists of two algorithms optimizing two objective functions defined by first-order proximity and second-order proximity. We prepared two link-generating models suitable for these two objective functions and clarified that the corresponding LINE algorithm performed well for the link data generated by each model.\",\"PeriodicalId\":444957,\"journal\":{\"name\":\"Proceedings of the 2022 5th International Conference on Software Engineering and Information Management\",\"volume\":\"82 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-01-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2022 5th International Conference on Software Engineering and Information Management\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3520084.3520111\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2022 5th International Conference on Software Engineering and Information Management","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3520084.3520111","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
An Evaluation of Large-scale Information Network Embedding based on Latent Space Model Generating Links
Graph representation learning encodes vertices as low-dimensional vectors that summarize their graph position and the structure of their local graph neighborhood. These methods give us beneficial representation in continuous space from big relational data. However, the algorithms are usually evaluated indirectly from the accuracy of applying the learning results to classification tasks because of not giving the correct answer when graph representation learning is applied. Therefore, this study proposes a method to evaluate graph representation learning algorithms by preparing correct learning results for the data by distributing objects in the latent space in advance and probabilistically generating relational graph data from the distributions in the latent space. Using this method, we evaluated LINE: Large-scale information network embedding, one of the most popular algorithms for learning graph representations. LINE consists of two algorithms optimizing two objective functions defined by first-order proximity and second-order proximity. We prepared two link-generating models suitable for these two objective functions and clarified that the corresponding LINE algorithm performed well for the link data generated by each model.