{"title":"High-speed convolutional neural networks using parallel prefix adders","authors":"Butchi Babu Sarnala, Siva Ramakrishna Pillutla","doi":"10.1016/j.aeue.2025.155791","DOIUrl":null,"url":null,"abstract":"<div><div>Many deep learning applications, particularly image processing, frequently utilize convolutional neural networks (CNNs) due to their remarkable accuracy. However, the heavy computational demands of CNNs pose a challenge for high-speed applications. First, this work introduces a delay efficient hybrid-parallel prefix adder (HPPA) where the total delay is minimized by changing black and gray cell paths. The proposed adder achieves lowest delay compared to the other available PPA adders. It achieves 10.64% reduction in total delay compared to recent PPA Adders. Secondly, we have employed the proposed HPPA adder to obtain a high-speed LeNet CNN architecture. Integration of the proposed hybrid adder into the LeNet CNN architecture significantly reduces the delay in the convolution layer. FPGA implementations of the proposed PPA adder and convolutional layer of LeNet architecture are performed. The proposed PPA adder achieves a 3.05% delay improvement while the critical path delay of the LeNet CNN is reduced by 16.78% compared to the relevant implementations. The implemented LeNet architecture using the proposed delay-efficient PPA adder offers a promising solution for deploying CNNs at high speed applications.</div></div>","PeriodicalId":50844,"journal":{"name":"Aeu-International Journal of Electronics and Communications","volume":"196 ","pages":"Article 155791"},"PeriodicalIF":3.0000,"publicationDate":"2025-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Aeu-International Journal of Electronics and Communications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1434841125001323","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Many deep learning applications, particularly image processing, frequently utilize convolutional neural networks (CNNs) due to their remarkable accuracy. However, the heavy computational demands of CNNs pose a challenge for high-speed applications. First, this work introduces a delay efficient hybrid-parallel prefix adder (HPPA) where the total delay is minimized by changing black and gray cell paths. The proposed adder achieves lowest delay compared to the other available PPA adders. It achieves 10.64% reduction in total delay compared to recent PPA Adders. Secondly, we have employed the proposed HPPA adder to obtain a high-speed LeNet CNN architecture. Integration of the proposed hybrid adder into the LeNet CNN architecture significantly reduces the delay in the convolution layer. FPGA implementations of the proposed PPA adder and convolutional layer of LeNet architecture are performed. The proposed PPA adder achieves a 3.05% delay improvement while the critical path delay of the LeNet CNN is reduced by 16.78% compared to the relevant implementations. The implemented LeNet architecture using the proposed delay-efficient PPA adder offers a promising solution for deploying CNNs at high speed applications.
期刊介绍:
AEÜ is an international scientific journal which publishes both original works and invited tutorials. The journal''s scope covers all aspects of theory and design of circuits, systems and devices for electronics, signal processing, and communication, including:
signal and system theory, digital signal processing
network theory and circuit design
information theory, communication theory and techniques, modulation, source and channel coding
switching theory and techniques, communication protocols
optical communications
microwave theory and techniques, radar, sonar
antennas, wave propagation
AEÜ publishes full papers and letters with very short turn around time but a high standard review process. Review cycles are typically finished within twelve weeks by application of modern electronic communication facilities.