参数相关矩阵的随机低阶近似

IF 1.8 3区 数学 Q1 MATHEMATICS
Daniel Kressner, Hei Yin Lam
{"title":"参数相关矩阵的随机低阶近似","authors":"Daniel Kressner, Hei Yin Lam","doi":"10.1002/nla.2576","DOIUrl":null,"url":null,"abstract":"This work considers the low‐rank approximation of a matrix depending on a parameter in a compact set . Application areas that give rise to such problems include computational statistics and dynamical systems. Randomized algorithms are an increasingly popular approach for performing low‐rank approximation and they usually proceed by multiplying the matrix with random dimension reduction matrices (DRMs). Applying such algorithms directly to would involve different, independent DRMs for every , which is not only expensive but also leads to inherently non‐smooth approximations. In this work, we propose to use constant DRMs, that is, is multiplied with the same DRM for every . The resulting parameter‐dependent extensions of two popular randomized algorithms, the randomized singular value decomposition and the generalized Nyström method, are computationally attractive, especially when admits an affine linear decomposition with respect to . We perform a probabilistic analysis for both algorithms, deriving bounds on the expected value as well as failure probabilities for the approximation error when using Gaussian random DRMs. Both, the theoretical results and numerical experiments, show that the use of constant DRMs does not impair their effectiveness; our methods reliably return quasi‐best low‐rank approximations.","PeriodicalId":49731,"journal":{"name":"Numerical Linear Algebra with Applications","volume":null,"pages":null},"PeriodicalIF":1.8000,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Randomized low‐rank approximation of parameter‐dependent matrices\",\"authors\":\"Daniel Kressner, Hei Yin Lam\",\"doi\":\"10.1002/nla.2576\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This work considers the low‐rank approximation of a matrix depending on a parameter in a compact set . Application areas that give rise to such problems include computational statistics and dynamical systems. Randomized algorithms are an increasingly popular approach for performing low‐rank approximation and they usually proceed by multiplying the matrix with random dimension reduction matrices (DRMs). Applying such algorithms directly to would involve different, independent DRMs for every , which is not only expensive but also leads to inherently non‐smooth approximations. In this work, we propose to use constant DRMs, that is, is multiplied with the same DRM for every . The resulting parameter‐dependent extensions of two popular randomized algorithms, the randomized singular value decomposition and the generalized Nyström method, are computationally attractive, especially when admits an affine linear decomposition with respect to . We perform a probabilistic analysis for both algorithms, deriving bounds on the expected value as well as failure probabilities for the approximation error when using Gaussian random DRMs. Both, the theoretical results and numerical experiments, show that the use of constant DRMs does not impair their effectiveness; our methods reliably return quasi‐best low‐rank approximations.\",\"PeriodicalId\":49731,\"journal\":{\"name\":\"Numerical Linear Algebra with Applications\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":1.8000,\"publicationDate\":\"2024-07-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Numerical Linear Algebra with Applications\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://doi.org/10.1002/nla.2576\",\"RegionNum\":3,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MATHEMATICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Numerical Linear Algebra with Applications","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1002/nla.2576","RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS","Score":null,"Total":0}
引用次数: 0

摘要

这项工作考虑的是在一个紧凑集合中对取决于参数的矩阵进行低秩逼近。引发此类问题的应用领域包括计算统计和动力系统。随机算法是进行低秩逼近的一种越来越流行的方法,它们通常通过将矩阵与随机降维矩阵(DRM)相乘来进行。直接应用这种算法将涉及每个秩矩阵的不同、独立 DRM,不仅成本高昂,而且会导致固有的非平滑近似。在这项工作中,我们建议使用恒定 DRM,也就是说,在每一个......乘以相同的 DRM。由此产生的两种流行随机算法--随机奇异值分解法和广义 Nyström 法--的参数扩展在计算上很有吸引力,尤其是当承认关于 .的仿射线性分解时。 我们对这两种算法进行了概率分析,推导出了使用高斯随机 DRM 时近似误差的期望值边界和失败概率。理论结果和数值实验都表明,使用常量 DRM 不会影响其有效性;我们的方法能可靠地返回准最佳低阶近似值。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Randomized low‐rank approximation of parameter‐dependent matrices
This work considers the low‐rank approximation of a matrix depending on a parameter in a compact set . Application areas that give rise to such problems include computational statistics and dynamical systems. Randomized algorithms are an increasingly popular approach for performing low‐rank approximation and they usually proceed by multiplying the matrix with random dimension reduction matrices (DRMs). Applying such algorithms directly to would involve different, independent DRMs for every , which is not only expensive but also leads to inherently non‐smooth approximations. In this work, we propose to use constant DRMs, that is, is multiplied with the same DRM for every . The resulting parameter‐dependent extensions of two popular randomized algorithms, the randomized singular value decomposition and the generalized Nyström method, are computationally attractive, especially when admits an affine linear decomposition with respect to . We perform a probabilistic analysis for both algorithms, deriving bounds on the expected value as well as failure probabilities for the approximation error when using Gaussian random DRMs. Both, the theoretical results and numerical experiments, show that the use of constant DRMs does not impair their effectiveness; our methods reliably return quasi‐best low‐rank approximations.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
3.40
自引率
2.30%
发文量
50
审稿时长
12 months
期刊介绍: Manuscripts submitted to Numerical Linear Algebra with Applications should include large-scale broad-interest applications in which challenging computational results are integral to the approach investigated and analysed. Manuscripts that, in the Editor’s view, do not satisfy these conditions will not be accepted for review. Numerical Linear Algebra with Applications receives submissions in areas that address developing, analysing and applying linear algebra algorithms for solving problems arising in multilinear (tensor) algebra, in statistics, such as Markov Chains, as well as in deterministic and stochastic modelling of large-scale networks, algorithm development, performance analysis or related computational aspects. Topics covered include: Standard and Generalized Conjugate Gradients, Multigrid and Other Iterative Methods; Preconditioning Methods; Direct Solution Methods; Numerical Methods for Eigenproblems; Newton-like Methods for Nonlinear Equations; Parallel and Vectorizable Algorithms in Numerical Linear Algebra; Application of Methods of Numerical Linear Algebra in Science, Engineering and Economics.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信