Dmitrii Khizbullin, Eduardo Rocha de Andrade, Thanh Hau Nguyen, Matheus Pedroza Ferreira, David R. Pugh
{"title":"为张量编译器配置交叉注意的图神经网络","authors":"Dmitrii Khizbullin, Eduardo Rocha de Andrade, Thanh Hau Nguyen, Matheus Pedroza Ferreira, David R. Pugh","doi":"arxiv-2405.16623","DOIUrl":null,"url":null,"abstract":"With the recent popularity of neural networks comes the need for efficient\nserving of inference workloads. A neural network inference workload can be\nrepresented as a computational graph with nodes as operators transforming\nmultidimensional tensors. The tensors can be transposed and/or tiled in a\ncombinatorially large number of ways, some configurations leading to\naccelerated inference. We propose TGraph, a neural graph architecture that\nallows screening for fast configurations of the target computational graph,\nthus representing an artificial intelligence (AI) tensor compiler in contrast\nto the traditional heuristics-based compilers. The proposed solution improves\nmean Kendall's $\\tau$ across layout collections of TpuGraphs from 29.8% of the\nreliable baseline to 67.4% of TGraph. We estimate the potential CO$_2$ emission\nreduction associated with our work to be equivalent to over 50% of the total\nhousehold emissions in the areas hosting AI-oriented data centers.","PeriodicalId":501291,"journal":{"name":"arXiv - CS - Performance","volume":"66 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Graph neural networks with configuration cross-attention for tensor compilers\",\"authors\":\"Dmitrii Khizbullin, Eduardo Rocha de Andrade, Thanh Hau Nguyen, Matheus Pedroza Ferreira, David R. Pugh\",\"doi\":\"arxiv-2405.16623\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With the recent popularity of neural networks comes the need for efficient\\nserving of inference workloads. A neural network inference workload can be\\nrepresented as a computational graph with nodes as operators transforming\\nmultidimensional tensors. The tensors can be transposed and/or tiled in a\\ncombinatorially large number of ways, some configurations leading to\\naccelerated inference. We propose TGraph, a neural graph architecture that\\nallows screening for fast configurations of the target computational graph,\\nthus representing an artificial intelligence (AI) tensor compiler in contrast\\nto the traditional heuristics-based compilers. The proposed solution improves\\nmean Kendall's $\\\\tau$ across layout collections of TpuGraphs from 29.8% of the\\nreliable baseline to 67.4% of TGraph. We estimate the potential CO$_2$ emission\\nreduction associated with our work to be equivalent to over 50% of the total\\nhousehold emissions in the areas hosting AI-oriented data centers.\",\"PeriodicalId\":501291,\"journal\":{\"name\":\"arXiv - CS - Performance\",\"volume\":\"66 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-05-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Performance\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2405.16623\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Performance","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.16623","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Graph neural networks with configuration cross-attention for tensor compilers
With the recent popularity of neural networks comes the need for efficient
serving of inference workloads. A neural network inference workload can be
represented as a computational graph with nodes as operators transforming
multidimensional tensors. The tensors can be transposed and/or tiled in a
combinatorially large number of ways, some configurations leading to
accelerated inference. We propose TGraph, a neural graph architecture that
allows screening for fast configurations of the target computational graph,
thus representing an artificial intelligence (AI) tensor compiler in contrast
to the traditional heuristics-based compilers. The proposed solution improves
mean Kendall's $\tau$ across layout collections of TpuGraphs from 29.8% of the
reliable baseline to 67.4% of TGraph. We estimate the potential CO$_2$ emission
reduction associated with our work to be equivalent to over 50% of the total
household emissions in the areas hosting AI-oriented data centers.