Jie-Fang Zhang, Ching-En Lee, Chester Liu, Y. Shao, S. Keckler, Zhengya Zhang
{"title":"SNAP:用于16nm CMOS非结构化稀疏深度神经网络推理的1.67 - 21.55TOPS/W稀疏神经加速处理器","authors":"Jie-Fang Zhang, Ching-En Lee, Chester Liu, Y. Shao, S. Keckler, Zhengya Zhang","doi":"10.23919/VLSIC.2019.8778193","DOIUrl":null,"url":null,"abstract":"A Sparse Neural Acceleration Processor (SNAP) is designed to exploit unstructured sparsity in deep neural networks (DNNs). SNAP uses parallel associative search to discover input pairs to maintain an average 75% hardware utilization. SNAP’s two-level partial sum reduce eliminates access contention and cuts the writeback traffic by $22\\times$. Through diagonal and row configurations of PE arrays, SNAP supports any CONV and FC layers. A 2.4mm2 16 nm SNAP test chip is measured to achieve a peak effectual efficiency of 21.55TOPS/W (16b) at 0.55V and 260MHz for CONV layers with 10% weight and activation density. Operating on pruned ResNet-50, SNAP achieves 90.98fps at 0.80V and 480MHz, dissipating 348mW.","PeriodicalId":6707,"journal":{"name":"2019 Symposium on VLSI Circuits","volume":"14 1","pages":"C306-C307"},"PeriodicalIF":0.0000,"publicationDate":"2019-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"35","resultStr":"{\"title\":\"SNAP: A 1.67 — 21.55TOPS/W Sparse Neural Acceleration Processor for Unstructured Sparse Deep Neural Network Inference in 16nm CMOS\",\"authors\":\"Jie-Fang Zhang, Ching-En Lee, Chester Liu, Y. Shao, S. Keckler, Zhengya Zhang\",\"doi\":\"10.23919/VLSIC.2019.8778193\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A Sparse Neural Acceleration Processor (SNAP) is designed to exploit unstructured sparsity in deep neural networks (DNNs). SNAP uses parallel associative search to discover input pairs to maintain an average 75% hardware utilization. SNAP’s two-level partial sum reduce eliminates access contention and cuts the writeback traffic by $22\\\\times$. Through diagonal and row configurations of PE arrays, SNAP supports any CONV and FC layers. A 2.4mm2 16 nm SNAP test chip is measured to achieve a peak effectual efficiency of 21.55TOPS/W (16b) at 0.55V and 260MHz for CONV layers with 10% weight and activation density. Operating on pruned ResNet-50, SNAP achieves 90.98fps at 0.80V and 480MHz, dissipating 348mW.\",\"PeriodicalId\":6707,\"journal\":{\"name\":\"2019 Symposium on VLSI Circuits\",\"volume\":\"14 1\",\"pages\":\"C306-C307\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-06-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"35\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 Symposium on VLSI Circuits\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.23919/VLSIC.2019.8778193\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 Symposium on VLSI Circuits","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.23919/VLSIC.2019.8778193","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
SNAP: A 1.67 — 21.55TOPS/W Sparse Neural Acceleration Processor for Unstructured Sparse Deep Neural Network Inference in 16nm CMOS
A Sparse Neural Acceleration Processor (SNAP) is designed to exploit unstructured sparsity in deep neural networks (DNNs). SNAP uses parallel associative search to discover input pairs to maintain an average 75% hardware utilization. SNAP’s two-level partial sum reduce eliminates access contention and cuts the writeback traffic by $22\times$. Through diagonal and row configurations of PE arrays, SNAP supports any CONV and FC layers. A 2.4mm2 16 nm SNAP test chip is measured to achieve a peak effectual efficiency of 21.55TOPS/W (16b) at 0.55V and 260MHz for CONV layers with 10% weight and activation density. Operating on pruned ResNet-50, SNAP achieves 90.98fps at 0.80V and 480MHz, dissipating 348mW.