Mahdi Ghorbani, Emilien Bauer, Tobias Grosser, Amir Shaikhha
{"title":"Compressing Structured Tensor Algebra","authors":"Mahdi Ghorbani, Emilien Bauer, Tobias Grosser, Amir Shaikhha","doi":"arxiv-2407.13726","DOIUrl":null,"url":null,"abstract":"Tensor algebra is a crucial component for data-intensive workloads such as\nmachine learning and scientific computing. As the complexity of data grows,\nscientists often encounter a dilemma between the highly specialized dense\ntensor algebra and efficient structure-aware algorithms provided by sparse\ntensor algebra. In this paper, we introduce DASTAC, a framework to propagate\nthe tensors's captured high-level structure down to low-level code generation\nby incorporating techniques such as automatic data layout compression,\npolyhedral analysis, and affine code generation. Our methodology reduces memory\nfootprint by automatically detecting the best data layout, heavily benefits\nfrom polyhedral optimizations, leverages further optimizations, and enables\nparallelization through MLIR. Through extensive experimentation, we show that\nDASTAC achieves 1 to 2 orders of magnitude speedup over TACO, a\nstate-of-the-art sparse tensor compiler, and StructTensor, a state-of-the-art\nstructured tensor algebra compiler, with a significantly lower memory\nfootprint.","PeriodicalId":501197,"journal":{"name":"arXiv - CS - Programming Languages","volume":"57 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Programming Languages","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.13726","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Tensor algebra is a crucial component for data-intensive workloads such as
machine learning and scientific computing. As the complexity of data grows,
scientists often encounter a dilemma between the highly specialized dense
tensor algebra and efficient structure-aware algorithms provided by sparse
tensor algebra. In this paper, we introduce DASTAC, a framework to propagate
the tensors's captured high-level structure down to low-level code generation
by incorporating techniques such as automatic data layout compression,
polyhedral analysis, and affine code generation. Our methodology reduces memory
footprint by automatically detecting the best data layout, heavily benefits
from polyhedral optimizations, leverages further optimizations, and enables
parallelization through MLIR. Through extensive experimentation, we show that
DASTAC achieves 1 to 2 orders of magnitude speedup over TACO, a
state-of-the-art sparse tensor compiler, and StructTensor, a state-of-the-art
structured tensor algebra compiler, with a significantly lower memory
footprint.