Feng Zhang, Xuebin Chi, Jinrong Jiang, Junlin Wei, Lian Zhao, Yuzhu Wang
{"title":"ilu预条件系统中多项式加速的一种有效标度方案","authors":"Feng Zhang, Xuebin Chi, Jinrong Jiang, Junlin Wei, Lian Zhao, Yuzhu Wang","doi":"10.1002/cpe.70098","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>Polynomial preconditioning accelerates iterative methods for large-scale sparse linear systems by optimizing the spectral distribution and decreasing reduction communication overhead. The Neumann polynomial is notable for its simple construction and stable performance, making it easy to combine with other preconditioners and widely used in high-performance computing. The choice of scaling parameter within the Neumann series is critical for polynomial acceleration, requiring an accurate estimate of the eigenvalue bounds of the preconditioned system. In preconditioned systems, the clustering of the largest eigenvalues often slows the convergence of iterative methods used to estimate the maximum eigenvalue, leading to an underestimated scaling parameter. We address this issue by using a Least-Squares model with linear inequality constraints to learn effective combination weights of Ritz values from training samples. While the Rayleigh-Ritz process (the current best eigen-estimation approach) requires 20-30 iterations and systematically underestimates extremal eigenvalues due to Ritz values' interior spectral distribution, our constrained optimization approach achieves comparable accuracy in 10 iterations by learning optimal combination weights from Ritz value distributions and corrects the systematic underestimation while preserving positive definitenessa critical stability requirement that ensures robust preconditioning performance across diverse problem configurations. Our implementation of the Neumann polynomial with the proposed scaling scheme achieved acceleration ratios of 2.61 and 3.52 for ILU (Incomplete LU factorization) and block-ILU preconditioned systems, respectively. It achieves comparable acceleration with the recent state-of-the-art minimum residual polynomial in the ILU-preconditioned systems frequently providing better convergence acceleration in numerous practical scenarios.</p>\n </div>","PeriodicalId":55214,"journal":{"name":"Concurrency and Computation-Practice & Experience","volume":"37 9-11","pages":""},"PeriodicalIF":1.5000,"publicationDate":"2025-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Efficient Scaling Scheme for Polynomial Acceleration in ILU-Preconditioned Systems\",\"authors\":\"Feng Zhang, Xuebin Chi, Jinrong Jiang, Junlin Wei, Lian Zhao, Yuzhu Wang\",\"doi\":\"10.1002/cpe.70098\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n <p>Polynomial preconditioning accelerates iterative methods for large-scale sparse linear systems by optimizing the spectral distribution and decreasing reduction communication overhead. The Neumann polynomial is notable for its simple construction and stable performance, making it easy to combine with other preconditioners and widely used in high-performance computing. The choice of scaling parameter within the Neumann series is critical for polynomial acceleration, requiring an accurate estimate of the eigenvalue bounds of the preconditioned system. In preconditioned systems, the clustering of the largest eigenvalues often slows the convergence of iterative methods used to estimate the maximum eigenvalue, leading to an underestimated scaling parameter. We address this issue by using a Least-Squares model with linear inequality constraints to learn effective combination weights of Ritz values from training samples. While the Rayleigh-Ritz process (the current best eigen-estimation approach) requires 20-30 iterations and systematically underestimates extremal eigenvalues due to Ritz values' interior spectral distribution, our constrained optimization approach achieves comparable accuracy in 10 iterations by learning optimal combination weights from Ritz value distributions and corrects the systematic underestimation while preserving positive definitenessa critical stability requirement that ensures robust preconditioning performance across diverse problem configurations. Our implementation of the Neumann polynomial with the proposed scaling scheme achieved acceleration ratios of 2.61 and 3.52 for ILU (Incomplete LU factorization) and block-ILU preconditioned systems, respectively. It achieves comparable acceleration with the recent state-of-the-art minimum residual polynomial in the ILU-preconditioned systems frequently providing better convergence acceleration in numerous practical scenarios.</p>\\n </div>\",\"PeriodicalId\":55214,\"journal\":{\"name\":\"Concurrency and Computation-Practice & Experience\",\"volume\":\"37 9-11\",\"pages\":\"\"},\"PeriodicalIF\":1.5000,\"publicationDate\":\"2025-04-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Concurrency and Computation-Practice & Experience\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/cpe.70098\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"COMPUTER SCIENCE, SOFTWARE ENGINEERING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Concurrency and Computation-Practice & Experience","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/cpe.70098","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
An Efficient Scaling Scheme for Polynomial Acceleration in ILU-Preconditioned Systems
Polynomial preconditioning accelerates iterative methods for large-scale sparse linear systems by optimizing the spectral distribution and decreasing reduction communication overhead. The Neumann polynomial is notable for its simple construction and stable performance, making it easy to combine with other preconditioners and widely used in high-performance computing. The choice of scaling parameter within the Neumann series is critical for polynomial acceleration, requiring an accurate estimate of the eigenvalue bounds of the preconditioned system. In preconditioned systems, the clustering of the largest eigenvalues often slows the convergence of iterative methods used to estimate the maximum eigenvalue, leading to an underestimated scaling parameter. We address this issue by using a Least-Squares model with linear inequality constraints to learn effective combination weights of Ritz values from training samples. While the Rayleigh-Ritz process (the current best eigen-estimation approach) requires 20-30 iterations and systematically underestimates extremal eigenvalues due to Ritz values' interior spectral distribution, our constrained optimization approach achieves comparable accuracy in 10 iterations by learning optimal combination weights from Ritz value distributions and corrects the systematic underestimation while preserving positive definitenessa critical stability requirement that ensures robust preconditioning performance across diverse problem configurations. Our implementation of the Neumann polynomial with the proposed scaling scheme achieved acceleration ratios of 2.61 and 3.52 for ILU (Incomplete LU factorization) and block-ILU preconditioned systems, respectively. It achieves comparable acceleration with the recent state-of-the-art minimum residual polynomial in the ILU-preconditioned systems frequently providing better convergence acceleration in numerous practical scenarios.
期刊介绍:
Concurrency and Computation: Practice and Experience (CCPE) publishes high-quality, original research papers, and authoritative research review papers, in the overlapping fields of:
Parallel and distributed computing;
High-performance computing;
Computational and data science;
Artificial intelligence and machine learning;
Big data applications, algorithms, and systems;
Network science;
Ontologies and semantics;
Security and privacy;
Cloud/edge/fog computing;
Green computing; and
Quantum computing.