投影张量-张量积用于最优多路数据表示的有效计算

IF 1.1 3区 数学 Q1 MATHEMATICS
Katherine Keegan , Elizabeth Newman
{"title":"投影张量-张量积用于最优多路数据表示的有效计算","authors":"Katherine Keegan ,&nbsp;Elizabeth Newman","doi":"10.1016/j.laa.2025.09.018","DOIUrl":null,"url":null,"abstract":"<div><div>Tensor decompositions have become essential tools for feature extraction and compression of multiway data. Recent advances in tensor operators have enabled desirable properties of standard matrix algebra to be retained for multilinear factorizations. Behind this matrix-mimetic tensor operation is an invertible matrix whose size depends quadratically on certain dimensions of the data. As a result, for large-scale multiway data, the invertible matrix can be computationally demanding to apply and invert and can lead to inefficient tensor representations in terms of construction and storage costs. In this work, we propose a new projected tensor-tensor product that relaxes the invertibility restriction to reduce computational overhead and still preserves fundamental linear algebraic properties. The transformation behind the projected product is a tall-and-skinny matrix with unitary columns, which depends only linearly on certain dimensions of the data, thereby reducing computational complexity by an order of magnitude. We provide extensive theory to prove the matrix mimeticity and the optimality of compressed representations within the projected product framework. We further prove that projected-product-based approximations outperform a comparable, non-matrix-mimetic tensor factorization. We support the theoretical findings and demonstrate the practical benefits of projected products through numerical experiments on video, hyperspectral imaging, synthetic, and dynamical systems data. All code for this paper is available at <span><span>https://github.com/elizabethnewman/projected-products.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18043,"journal":{"name":"Linear Algebra and its Applications","volume":"729 ","pages":"Pages 100-147"},"PeriodicalIF":1.1000,"publicationDate":"2025-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Projected tensor-tensor products for efficient computation of optimal multiway data representations\",\"authors\":\"Katherine Keegan ,&nbsp;Elizabeth Newman\",\"doi\":\"10.1016/j.laa.2025.09.018\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Tensor decompositions have become essential tools for feature extraction and compression of multiway data. Recent advances in tensor operators have enabled desirable properties of standard matrix algebra to be retained for multilinear factorizations. Behind this matrix-mimetic tensor operation is an invertible matrix whose size depends quadratically on certain dimensions of the data. As a result, for large-scale multiway data, the invertible matrix can be computationally demanding to apply and invert and can lead to inefficient tensor representations in terms of construction and storage costs. In this work, we propose a new projected tensor-tensor product that relaxes the invertibility restriction to reduce computational overhead and still preserves fundamental linear algebraic properties. The transformation behind the projected product is a tall-and-skinny matrix with unitary columns, which depends only linearly on certain dimensions of the data, thereby reducing computational complexity by an order of magnitude. We provide extensive theory to prove the matrix mimeticity and the optimality of compressed representations within the projected product framework. We further prove that projected-product-based approximations outperform a comparable, non-matrix-mimetic tensor factorization. We support the theoretical findings and demonstrate the practical benefits of projected products through numerical experiments on video, hyperspectral imaging, synthetic, and dynamical systems data. All code for this paper is available at <span><span>https://github.com/elizabethnewman/projected-products.git</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":18043,\"journal\":{\"name\":\"Linear Algebra and its Applications\",\"volume\":\"729 \",\"pages\":\"Pages 100-147\"},\"PeriodicalIF\":1.1000,\"publicationDate\":\"2025-09-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Linear Algebra and its Applications\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S002437952500391X\",\"RegionNum\":3,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MATHEMATICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Linear Algebra and its Applications","FirstCategoryId":"100","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S002437952500391X","RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS","Score":null,"Total":0}
引用次数: 0

摘要

张量分解已经成为多路数据特征提取和压缩的重要工具。张量算子的最新进展使标准矩阵代数的理想性质在多线性分解中得以保留。在这个矩阵模拟张量运算的背后是一个可逆矩阵,它的大小二次依赖于数据的某些维度。因此,对于大规模的多路数据,可逆矩阵的应用和反转可能在计算上要求很高,并且在构造和存储成本方面可能导致低效的张量表示。在这项工作中,我们提出了一个新的投影张量-张量积,它放宽了可逆性限制,以减少计算开销,并且仍然保持基本的线性代数性质。投影乘积背后的变换是一个具有酉列的又高又瘦的矩阵,它仅线性地依赖于数据的某些维度,从而将计算复杂性降低了一个数量级。我们提供了广泛的理论来证明矩阵拟性和在投影产品框架内压缩表示的最优性。我们进一步证明了基于投影积的近似优于可比的非矩阵模拟张量分解。我们支持理论发现,并通过视频、高光谱成像、合成和动力系统数据的数值实验证明了投影产品的实际效益。本文的所有代码可在https://github.com/elizabethnewman/projected-products.git上获得。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Projected tensor-tensor products for efficient computation of optimal multiway data representations
Tensor decompositions have become essential tools for feature extraction and compression of multiway data. Recent advances in tensor operators have enabled desirable properties of standard matrix algebra to be retained for multilinear factorizations. Behind this matrix-mimetic tensor operation is an invertible matrix whose size depends quadratically on certain dimensions of the data. As a result, for large-scale multiway data, the invertible matrix can be computationally demanding to apply and invert and can lead to inefficient tensor representations in terms of construction and storage costs. In this work, we propose a new projected tensor-tensor product that relaxes the invertibility restriction to reduce computational overhead and still preserves fundamental linear algebraic properties. The transformation behind the projected product is a tall-and-skinny matrix with unitary columns, which depends only linearly on certain dimensions of the data, thereby reducing computational complexity by an order of magnitude. We provide extensive theory to prove the matrix mimeticity and the optimality of compressed representations within the projected product framework. We further prove that projected-product-based approximations outperform a comparable, non-matrix-mimetic tensor factorization. We support the theoretical findings and demonstrate the practical benefits of projected products through numerical experiments on video, hyperspectral imaging, synthetic, and dynamical systems data. All code for this paper is available at https://github.com/elizabethnewman/projected-products.git.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
2.20
自引率
9.10%
发文量
333
审稿时长
13.8 months
期刊介绍: Linear Algebra and its Applications publishes articles that contribute new information or new insights to matrix theory and finite dimensional linear algebra in their algebraic, arithmetic, combinatorial, geometric, or numerical aspects. It also publishes articles that give significant applications of matrix theory or linear algebra to other branches of mathematics and to other sciences. Articles that provide new information or perspectives on the historical development of matrix theory and linear algebra are also welcome. Expository articles which can serve as an introduction to a subject for workers in related areas and which bring one to the frontiers of research are encouraged. Reviews of books are published occasionally as are conference reports that provide an historical record of major meetings on matrix theory and linear algebra.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信