Zhao-Li Shen , Guo-Liang Han , Yu-Tong Liu , Bruno Carpentieri , Chun Wen , Jian-Jun Wang
{"title":"Weak dangling block reordering and multi-step block compression for efficiently computing and updating PageRank solutions","authors":"Zhao-Li Shen , Guo-Liang Han , Yu-Tong Liu , Bruno Carpentieri , Chun Wen , Jian-Jun Wang","doi":"10.1016/j.cam.2024.116332","DOIUrl":null,"url":null,"abstract":"<div><div>The PageRank model is a powerful tool for network analysis, utilized across various disciplines such as web information retrieval, bioinformatics, community detection, and graph neural network. Computing this model requires solving a large-dimensional linear system or eigenvector problem due to the ever-increasing scale of networks. Conventional preconditioners and iterative methods for general linear systems or eigenvector problems often exhibit unsatisfactory performance for such problems, particularly as the damping factor parameter approaches 1, necessitating the development of specialized methods that exploit the specific properties of the PageRank coefficient matrix. Additionally, in practical applications, the optimal settings of the hyperparameters are generally unknown in advance, and networks often evolve over time. Consequently, recomputation of the problem is necessary following minor modifications. In this scenario, highly efficient preconditioners that significantly accelerate the iterative solution at a low memory cost are desirable. In this paper, we present two techniques that leverage the sparsity structures and numerical properties of the PageRank system, as well as a preconditioner based on the computed matrix structure. Experiments demonstrate the positive performance of the proposed methods on realistic PageRank computations.</div></div>","PeriodicalId":50226,"journal":{"name":"Journal of Computational and Applied Mathematics","volume":"458 ","pages":"Article 116332"},"PeriodicalIF":2.1000,"publicationDate":"2024-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Computational and Applied Mathematics","FirstCategoryId":"100","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0377042724005806","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS, APPLIED","Score":null,"Total":0}
引用次数: 0
Abstract
The PageRank model is a powerful tool for network analysis, utilized across various disciplines such as web information retrieval, bioinformatics, community detection, and graph neural network. Computing this model requires solving a large-dimensional linear system or eigenvector problem due to the ever-increasing scale of networks. Conventional preconditioners and iterative methods for general linear systems or eigenvector problems often exhibit unsatisfactory performance for such problems, particularly as the damping factor parameter approaches 1, necessitating the development of specialized methods that exploit the specific properties of the PageRank coefficient matrix. Additionally, in practical applications, the optimal settings of the hyperparameters are generally unknown in advance, and networks often evolve over time. Consequently, recomputation of the problem is necessary following minor modifications. In this scenario, highly efficient preconditioners that significantly accelerate the iterative solution at a low memory cost are desirable. In this paper, we present two techniques that leverage the sparsity structures and numerical properties of the PageRank system, as well as a preconditioner based on the computed matrix structure. Experiments demonstrate the positive performance of the proposed methods on realistic PageRank computations.
期刊介绍:
The Journal of Computational and Applied Mathematics publishes original papers of high scientific value in all areas of computational and applied mathematics. The main interest of the Journal is in papers that describe and analyze new computational techniques for solving scientific or engineering problems. Also the improved analysis, including the effectiveness and applicability, of existing methods and algorithms is of importance. The computational efficiency (e.g. the convergence, stability, accuracy, ...) should be proved and illustrated by nontrivial numerical examples. Papers describing only variants of existing methods, without adding significant new computational properties are not of interest.
The audience consists of: applied mathematicians, numerical analysts, computational scientists and engineers.