Dmitrii Khizbullin, Eduardo Rocha de Andrade, Thanh Hau Nguyen, Matheus Pedroza Ferreira, David R. Pugh
{"title":"Graph neural networks with configuration cross-attention for tensor compilers","authors":"Dmitrii Khizbullin, Eduardo Rocha de Andrade, Thanh Hau Nguyen, Matheus Pedroza Ferreira, David R. Pugh","doi":"arxiv-2405.16623","DOIUrl":null,"url":null,"abstract":"With the recent popularity of neural networks comes the need for efficient\nserving of inference workloads. A neural network inference workload can be\nrepresented as a computational graph with nodes as operators transforming\nmultidimensional tensors. The tensors can be transposed and/or tiled in a\ncombinatorially large number of ways, some configurations leading to\naccelerated inference. We propose TGraph, a neural graph architecture that\nallows screening for fast configurations of the target computational graph,\nthus representing an artificial intelligence (AI) tensor compiler in contrast\nto the traditional heuristics-based compilers. The proposed solution improves\nmean Kendall's $\\tau$ across layout collections of TpuGraphs from 29.8% of the\nreliable baseline to 67.4% of TGraph. We estimate the potential CO$_2$ emission\nreduction associated with our work to be equivalent to over 50% of the total\nhousehold emissions in the areas hosting AI-oriented data centers.","PeriodicalId":501291,"journal":{"name":"arXiv - CS - Performance","volume":"66 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Performance","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.16623","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
With the recent popularity of neural networks comes the need for efficient
serving of inference workloads. A neural network inference workload can be
represented as a computational graph with nodes as operators transforming
multidimensional tensors. The tensors can be transposed and/or tiled in a
combinatorially large number of ways, some configurations leading to
accelerated inference. We propose TGraph, a neural graph architecture that
allows screening for fast configurations of the target computational graph,
thus representing an artificial intelligence (AI) tensor compiler in contrast
to the traditional heuristics-based compilers. The proposed solution improves
mean Kendall's $\tau$ across layout collections of TpuGraphs from 29.8% of the
reliable baseline to 67.4% of TGraph. We estimate the potential CO$_2$ emission
reduction associated with our work to be equivalent to over 50% of the total
household emissions in the areas hosting AI-oriented data centers.