Kyle Deeds, Willow Ahrens, Magda Balazinska, Dan Suciu
{"title":"Galley:稀疏张量程序的现代查询优化","authors":"Kyle Deeds, Willow Ahrens, Magda Balazinska, Dan Suciu","doi":"arxiv-2408.14706","DOIUrl":null,"url":null,"abstract":"The tensor programming abstraction has become the key . This framework allows\nusers to write high performance programs for bulk computation via a high-level\nimperative interface. Recent work has extended this paradigm to sparse tensors\n(i.e. tensors where most entries are not explicitly represented) with the use\nof sparse tensor compilers. These systems excel at producing efficient code for\ncomputation over sparse tensors, which may be stored in a wide variety of\nformats. However, they require the user to manually choose the order of\noperations and the data formats at every step. Unfortunately, these decisions\nare both highly impactful and complicated, requiring significant effort to\nmanually optimize. In this work, we present Galley, a system for declarative\nsparse tensor programming. Galley performs cost-based optimization to lower\nthese programs to a logical plan then to a physical plan. It then leverages\nsparse tensor compilers to execute the physical plan efficiently. We show that\nGalley achieves high performance on a wide variety of problems including\nmachine learning algorithms, subgraph counting, and iterative graph algorithms.","PeriodicalId":501197,"journal":{"name":"arXiv - CS - Programming Languages","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Galley: Modern Query Optimization for Sparse Tensor Programs\",\"authors\":\"Kyle Deeds, Willow Ahrens, Magda Balazinska, Dan Suciu\",\"doi\":\"arxiv-2408.14706\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The tensor programming abstraction has become the key . This framework allows\\nusers to write high performance programs for bulk computation via a high-level\\nimperative interface. Recent work has extended this paradigm to sparse tensors\\n(i.e. tensors where most entries are not explicitly represented) with the use\\nof sparse tensor compilers. These systems excel at producing efficient code for\\ncomputation over sparse tensors, which may be stored in a wide variety of\\nformats. However, they require the user to manually choose the order of\\noperations and the data formats at every step. Unfortunately, these decisions\\nare both highly impactful and complicated, requiring significant effort to\\nmanually optimize. In this work, we present Galley, a system for declarative\\nsparse tensor programming. Galley performs cost-based optimization to lower\\nthese programs to a logical plan then to a physical plan. It then leverages\\nsparse tensor compilers to execute the physical plan efficiently. We show that\\nGalley achieves high performance on a wide variety of problems including\\nmachine learning algorithms, subgraph counting, and iterative graph algorithms.\",\"PeriodicalId\":501197,\"journal\":{\"name\":\"arXiv - CS - Programming Languages\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Programming Languages\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2408.14706\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Programming Languages","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.14706","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Galley: Modern Query Optimization for Sparse Tensor Programs
The tensor programming abstraction has become the key . This framework allows
users to write high performance programs for bulk computation via a high-level
imperative interface. Recent work has extended this paradigm to sparse tensors
(i.e. tensors where most entries are not explicitly represented) with the use
of sparse tensor compilers. These systems excel at producing efficient code for
computation over sparse tensors, which may be stored in a wide variety of
formats. However, they require the user to manually choose the order of
operations and the data formats at every step. Unfortunately, these decisions
are both highly impactful and complicated, requiring significant effort to
manually optimize. In this work, we present Galley, a system for declarative
sparse tensor programming. Galley performs cost-based optimization to lower
these programs to a logical plan then to a physical plan. It then leverages
sparse tensor compilers to execute the physical plan efficiently. We show that
Galley achieves high performance on a wide variety of problems including
machine learning algorithms, subgraph counting, and iterative graph algorithms.