Analyzing different high speed adder architecture for Neural Networks

Deekshith Krishnegowda
{"title":"Analyzing different high speed adder architecture for Neural Networks","authors":"Deekshith Krishnegowda","doi":"10.1109/ICAECC54045.2022.9716643","DOIUrl":null,"url":null,"abstract":"The first neural network model which was developed for image recognition application consisted of simple perceptrons. It had input, processing unit, and a single output. Neural networks which are used in today’s world consist of many complex MAC (Multiply and Accumulate) units. Be it the simple pattern recognition neural network model or complex models used for autonomous driving applications; adders are used for computing the activation point of neurons. Some adders offer better performance at the cost of area and power while some offer better power at the cost of performance. So, choosing the right type of adder architecture based upon the application becomes a very important criterion when we are trying to develop an inference engine for the neural network in hardware. To determine weight or activation point of a neuron, typically, float32 or float64 number representation is used. Float64 offers better accuracy than float32 but the drawback of using float64 is that it requires huge computation power. So, in this manuscript we compare different high-speed adder topologies, then discuss the implementation of an optimized 64-bit conditional sum and carry select adder that can be used to implement Deep Neural Network with float64 number representation. Analysis between different adder architecture is performed using Synopsys Design Compiler with 45nm Toshiba library for three different metrics: Timing, Area, and Power.","PeriodicalId":199351,"journal":{"name":"2022 IEEE Fourth International Conference on Advances in Electronics, Computers and Communications (ICAECC)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE Fourth International Conference on Advances in Electronics, Computers and Communications (ICAECC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAECC54045.2022.9716643","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

The first neural network model which was developed for image recognition application consisted of simple perceptrons. It had input, processing unit, and a single output. Neural networks which are used in today’s world consist of many complex MAC (Multiply and Accumulate) units. Be it the simple pattern recognition neural network model or complex models used for autonomous driving applications; adders are used for computing the activation point of neurons. Some adders offer better performance at the cost of area and power while some offer better power at the cost of performance. So, choosing the right type of adder architecture based upon the application becomes a very important criterion when we are trying to develop an inference engine for the neural network in hardware. To determine weight or activation point of a neuron, typically, float32 or float64 number representation is used. Float64 offers better accuracy than float32 but the drawback of using float64 is that it requires huge computation power. So, in this manuscript we compare different high-speed adder topologies, then discuss the implementation of an optimized 64-bit conditional sum and carry select adder that can be used to implement Deep Neural Network with float64 number representation. Analysis between different adder architecture is performed using Synopsys Design Compiler with 45nm Toshiba library for three different metrics: Timing, Area, and Power.
神经网络中不同高速加法器结构分析
第一个用于图像识别应用的神经网络模型是由简单的感知器组成的。它有输入、处理单元和一个输出。当今世界上使用的神经网络由许多复杂的MAC(乘法和累加)单元组成。无论是简单的模式识别神经网络模型还是用于自动驾驶应用的复杂模型;加法器用于计算神经元的激活点。有些加法器以面积和功率为代价提供更好的性能,而有些则以性能为代价提供更好的功率。因此,在硬件上为神经网络开发推理引擎时,根据应用选择合适的加法器结构就成为一个非常重要的标准。为了确定神经元的权重或激活点,通常使用float32或float64数字表示。Float64提供比float32更好的精度,但使用Float64的缺点是它需要巨大的计算能力。因此,在本文中,我们比较了不同的高速加法器拓扑,然后讨论了一个优化的64位条件和进位选择加法器的实现,该加法器可用于实现具有float64数字表示的深度神经网络。不同加法器架构之间的分析使用Synopsys设计编译器与45纳米东芝库进行三个不同的指标:时序,面积和功耗。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信