ACM Transactions on Embedded Computing Systems最新文献

筛选
英文 中文
LiteHash: Hash Functions for Resource-Constrained Hardware LiteHash:适用于资源受限硬件的哈希函数
IF 2.8 3区 计算机科学
ACM Transactions on Embedded Computing Systems Pub Date : 2024-07-09 DOI: 10.1145/3677181
Sagar Dev Achar, Thejaswini P, Sukumar Nandi, S. Nandi
{"title":"LiteHash: Hash Functions for Resource-Constrained Hardware","authors":"Sagar Dev Achar, Thejaswini P, Sukumar Nandi, S. Nandi","doi":"10.1145/3677181","DOIUrl":"https://doi.org/10.1145/3677181","url":null,"abstract":"\u0000 The global paradigm shift towards edge computing has led to a growing demand for efficient integrity verification. Hash functions are one-way algorithms which act as a zero-knowledge proof of a datum’s contents. However, it is infeasible to compute hashes on devices with limited processing power and memory. Hence, we propose four novel\u0000 LiteHash\u0000 functions which are architecturally similar to SHA-512 yet simpler. By using various approximation techniques, our implementations reduce the computational costs of digesting a message into a hash. On validating our proposed designs using the NIST PRNG Test Suite, we observe SHA-512 equivalent cryptographic security while satisfying all desired hash function property requirements. We observe a minimum of 9.41% reduction in area, 20.47% reduction in power and 22.05% increase in throughput. Our designs offer a throughput of upto 2 Gbps while reducing area and power by a maximum of 16.86% and 32.48% respectively.\u0000 LiteHash\u0000 functions also support the computation of the entire SHA-2 family of hash functions (SHA-224/256/384/512) with minor architectural modifications.\u0000","PeriodicalId":50914,"journal":{"name":"ACM Transactions on Embedded Computing Systems","volume":null,"pages":null},"PeriodicalIF":2.8,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141663843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Hybrid Sparse-dense Defensive DNN Accelerator Architecture against Adversarial Example Attacks 针对对抗性示例攻击的稀疏-密集混合防御 DNN 加速体系结构
IF 2.8 3区 计算机科学
ACM Transactions on Embedded Computing Systems Pub Date : 2024-07-09 DOI: 10.1145/3677318
Xingbin Wang, Boyan Zhao, Yulan Su, Sisi Zhang, Fengkai Yuan, Jun Zhang, Dan Meng, Rui Hou
{"title":"A Hybrid Sparse-dense Defensive DNN Accelerator Architecture against Adversarial Example Attacks","authors":"Xingbin Wang, Boyan Zhao, Yulan Su, Sisi Zhang, Fengkai Yuan, Jun Zhang, Dan Meng, Rui Hou","doi":"10.1145/3677318","DOIUrl":"https://doi.org/10.1145/3677318","url":null,"abstract":"\u0000 Understanding how to defend against adversarial attacks is crucial for ensuring the safety and reliability of these systems in real-world applications. Various adversarial defense methods are proposed, which aim to improve the robustness of neural networks against adversarial attacks by changing the model structure, adding detection networks, and adversarial purification network. However, deploying adversarial defense methods in existing DNN accelerators or defensive accelerators leads to many key issues. To address these challenges, this paper proposes\u0000 sDNNGuard\u0000 , an elastic heterogeneous DNN accelerator architecture that can efficiently orchestrate the simultaneous execution of original (\u0000 target\u0000 ) DNN networks and the\u0000 detect\u0000 algorithm or network. It not only supports for dense DNN detect algorithms, but also allows for sparse DNN defense methods and other mixed dense-sparse (e.g., dense-dense and sparse-dense) workloads to fully exploit the benefits of sparsity. sDNNGuard with a CPU core also supports the non-DNN computing and allows the special layer of the neural network, and used for the conversion for sparse storage format for weights and activation values. To reduce off-chip traffic and improve resources utilization, a new hardware abstraction with elastic on-chip buffer/computing resource management is proposed to achieve dynamical resource scheduling mechanism. We propose an\u0000 extended AI instruction set\u0000 for neural networks synchronization, task scheduling and efficient data interaction. Experiment results show that sDNNGuard can effectively validate the legitimacy of the input samples in parallel with the target DNN model, achieving an average 1.42 × speedup compared with the state-of-the-art accelerators.\u0000","PeriodicalId":50914,"journal":{"name":"ACM Transactions on Embedded Computing Systems","volume":null,"pages":null},"PeriodicalIF":2.8,"publicationDate":"2024-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141663892","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
NIR-sighted: A Programmable Streaming Architecture for Low-Energy Human-Centric Vision Applications 近视:面向以人为本的低能耗视觉应用的可编程流架构
IF 2 3区 计算机科学
ACM Transactions on Embedded Computing Systems Pub Date : 2024-06-14 DOI: 10.1145/3672076
John Mamish, Rawan Alharbi, Sougata Sen, Shashank Holla, Panchami Kamath, Yaman Sangar, Nabil I Alshurafa, Josiah D. Hester
{"title":"NIR-sighted: A Programmable Streaming Architecture for Low-Energy Human-Centric Vision Applications","authors":"John Mamish, Rawan Alharbi, Sougata Sen, Shashank Holla, Panchami Kamath, Yaman Sangar, Nabil I Alshurafa, Josiah D. Hester","doi":"10.1145/3672076","DOIUrl":"https://doi.org/10.1145/3672076","url":null,"abstract":"Human studies often rely on wearable lifelogging cameras that capture videos of individuals and their surroundings to aid in visual confirmation or recollection of daily activities like eating, drinking and smoking. However, this may include private or sensitive information that may cause some users to refrain from using such monitoring devices. Also, short battery lifetime and large form factors reduce applicability for long-term capture of human activity. Solving this triad of interconnected problems is challenging due to wearable embedded systems’ energy, memory and computing constraints. Inspired by this critical use case and the unique design problem, we developed NIR-sighted, an architecture for wearable video cameras which navigates this design space via three key ideas: i)Reduce storage and enhance privacy by discarding masked pixels and frames. ii) Enable programmers to generate effective masks with low computational overhead. iii) Enable the use of small MCUs by moving masking and compression off-chip. Combined together in an end-to-end system, NIR-sighted’s masking capabilities and off-chip compression hardware shrinks systems, stores less data, and enables programmer-defined obfuscation to yield privacy enhancement. The user’s privacy is enhanced significantly as nowhere in the pipeline is any part of the image stored before it is obfuscated. We design a wearable camera called NIR-sightedCam based on this architecture; it is compact and can record IR and grayscale video at 16 and 20+fps respectively for 26 hours nonstop (59 hours with IR disabled) at a fraction of comparable platforms power draw. NIR-sightedCam includes a low-power FPGA which implements our mJPEG compress/obfuscate hardware, Blindspot. We additionally show the potential for privacy-enhancing function and clinical utility via an in-lab eating study, validated by a nutritionist.","PeriodicalId":50914,"journal":{"name":"ACM Transactions on Embedded Computing Systems","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141342686","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Traffic Resource Optimization for Real-Time Applications with 5G Configured Grant Scheduling 利用 5G 配置授权调度为实时应用优化多流量资源
IF 2 3区 计算机科学
ACM Transactions on Embedded Computing Systems Pub Date : 2024-05-28 DOI: 10.1145/3664621
Yungang Pan, Rouhollah Mahfouzi, Soheil Samii, Petru Eles, Zebo Peng
{"title":"Multi-Traffic Resource Optimization for Real-Time Applications with 5G Configured Grant Scheduling","authors":"Yungang Pan, Rouhollah Mahfouzi, Soheil Samii, Petru Eles, Zebo Peng","doi":"10.1145/3664621","DOIUrl":"https://doi.org/10.1145/3664621","url":null,"abstract":"<p>The fifth-generation (5G) technology standard in telecommunications is expected to support ultra-reliable low latency communication to enable real-time applications such as industrial automation and control. 5G configured grant (CG) scheduling features a pre-allocated periodicity-based scheduling approach, which reduces control signaling time and guarantees service quality. Although this enables 5G to support hard real-time periodic traffics, synthesizing the schedule efficiently and achieving high resource efficiency, while serving multiple communications, are still an open problem. In this work, we study the trade-off between scheduling flexibility and control overhead when performing CG scheduling. To address the CG scheduling problem, we first formulate it using satisfiability modulo theories (SMT) so that an SMT solver can be used to generate optimal solutions. To enhance scalability, we propose two heuristic approaches. The first one as the baseline, Co1, follows the basic idea of the 5G CG scheduling scheme that minimizes the control overhead. The second one, CoU, enables increased scheduling flexibility while considering the involved control overhead. The effectiveness and scalability of the proposed techniques and the superiority of CoU compared to Co1 have been evaluated using a large number of generated benchmarks as well as a realistic case study for industrial automation.</p>","PeriodicalId":50914,"journal":{"name":"ACM Transactions on Embedded Computing Systems","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141169425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic Cluster Head Selection in WSN WSN 中的动态簇头选择
IF 2 3区 计算机科学
ACM Transactions on Embedded Computing Systems Pub Date : 2024-05-25 DOI: 10.1145/3665867
Rupendra Pratap Singh Hada, Abhishek Srivastava
{"title":"Dynamic Cluster Head Selection in WSN","authors":"Rupendra Pratap Singh Hada, Abhishek Srivastava","doi":"10.1145/3665867","DOIUrl":"https://doi.org/10.1145/3665867","url":null,"abstract":"<p>A Wireless Sensor Network (WSN) comprises an ad-hoc network of nodes laden with sensors that are used to monitor a region mostly in the outdoors and often not easily accessible. Despite exceptions, several deployments of WSN continue to grapple with the limitation of finite energy derived through batteries. Thus, it is imperative that the energy of a WSN be conserved and its life prolonged. An important direction of work to this end is towards the transmission of data between nodes in a manner that minimum energy is expended. One approach to doing this is cluster-based routing, wherein nodes in a WSN are organised into clusters, and transmission of data from the node is through a representative node called a cluster-head. Forming optimal clusters and choosing an optimal cluster-head is an NP-Hard problem. Significant work is done towards devising mechanisms to form clusters and choosing cluster heads to reduce the transmission overhead to a minimum. In this paper, an approach is proposed to create clusters and identify cluster heads that are near optimal. The approach involves two-stage clustering, with the clustering algorithm for each stage chosen through an exhaustive search. Furthermore, unlike existing approaches that choose a cluster-head on the basis of the residual energy of nodes, the proposed approach utilises three factors in addition to the residual energy, namely the distance of a node from the cluster centroid, the distance of a node from the final destination (base-station), and the connectivity of the node. The approach is shown to be effective and economical through extensive validation via simulations and through a real-world prototypical implementation.</p>","PeriodicalId":50914,"journal":{"name":"ACM Transactions on Embedded Computing Systems","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141149096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CARIn: Constraint-Aware and Responsive Inference on Heterogeneous Devices for Single- and Multi-DNN Workloads CARIn:在异构设备上针对单 DNN 和多 DNN 工作负载进行约束感知和响应式推理
IF 2 3区 计算机科学
ACM Transactions on Embedded Computing Systems Pub Date : 2024-05-23 DOI: 10.1145/3665868
Ioannis Panopoulos, Stylianos I. Venieris, I. Venieris
{"title":"CARIn: Constraint-Aware and Responsive Inference on Heterogeneous Devices for Single- and Multi-DNN Workloads","authors":"Ioannis Panopoulos, Stylianos I. Venieris, I. Venieris","doi":"10.1145/3665868","DOIUrl":"https://doi.org/10.1145/3665868","url":null,"abstract":"\u0000 The relentless expansion of deep learning (DL) applications in recent years has prompted a pivotal shift towards on-device execution, driven by the urgent need for real-time processing, heightened privacy concerns, and reduced latency across diverse domains. This paper addresses the challenges inherent in optimising the execution of deep neural networks (DNNs) on mobile devices, with a focus on device heterogeneity, multi-DNN execution, and dynamic runtime adaptation. We introduce\u0000 CARIn\u0000 , a novel framework designed for the optimised deployment of both single- and multi-DNN applications under user-defined service-level objectives (SLOs). Leveraging an expressive multi-objective optimisation (MOO) framework and a runtime-aware sorting and search algorithm (\u0000 RASS\u0000 ) as the MOO solver,\u0000 CARIn\u0000 facilitates efficient adaptation to dynamic conditions while addressing resource contention issues associated with multi-DNN execution. Notably,\u0000 RASS\u0000 generates a set of configurations, anticipating subsequent runtime adaptation, ensuring rapid, low-overhead adjustments in response to environmental fluctuations. Extensive evaluation across diverse tasks, including text classification, scene recognition, and face analysis, showcases the versatility of\u0000 CARIn\u0000 across various model architectures, such as Convolutional Neural Networks (CNNs) and Transformers, and realistic use cases. We observe a substantial enhancement in the fair treatment of the problem’s objectives, reaching 1.92 × when compared to single-model designs, and up to 10.69 × in contrast to the state-of-the-art OODIn framework. Additionally, we achieve a significant gain of up to 4.06 × over hardware-unaware designs in multi-DNN applications. Finally, our framework sustains its performance while effectively eliminating the time overhead associated with identifying the optimal design in response to environmental challenges.\u0000","PeriodicalId":50914,"journal":{"name":"ACM Transactions on Embedded Computing Systems","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141105549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Hardware Approach For Accelerating Inductive Learning In Description Logic 加速描述逻辑归纳学习的硬件方法
IF 2 3区 计算机科学
ACM Transactions on Embedded Computing Systems Pub Date : 2024-05-21 DOI: 10.1145/3665277
Eyad Algahtani
{"title":"A Hardware Approach For Accelerating Inductive Learning In Description Logic","authors":"Eyad Algahtani","doi":"10.1145/3665277","DOIUrl":"https://doi.org/10.1145/3665277","url":null,"abstract":"\u0000 The employment of machine learning (ML) techniques in embedded systems, has seen constant growth in recent years, especially for black-box ML techniques (such as artificial neural networks, ANNs). However, despite the successful employment of ML techniques in embedded environments, yet, their performance potential is constrained by the limited computing resources of their embedded computers. Several hardware based approaches were developed (e.g. using FPGAs and ASICs), to address the constraints of limited computing resources. The scope of this work, focuses on improving the performance for Inductive Logic Programming (ILP) on embedded environments. ILP is a powerful logic-based ML technique that uses logic programming, to construct human-interpretable ML models; where those logic-based ML models, are capable of describing complex and multi-relational concepts. In this work, we present a hardware-based approach that accelerate the hypothesis evaluation task for ILPs in embedded environments, that uses Description Logic (DL) languages as their logic-based representation; In particular, we target the\u0000 \u0000 (mathcal {ALCQ}^{mathcal {(D)}} )\u0000 \u0000 language. According to experimental results (through an FPGA implementation), our presented approach has achieved speedups up to 48.7 folds for a disjunction of 32 concepts on 100M individuals; where the baseline performance is the sequential CPU performance of the Raspberry Pi 4. For role and concrete role restrictions, the FPGA implementation achieved speedups up to 2.4 folds (for MIN cardinality role restriction on 1M role assertions); all FPGA implemented role and concrete role restrictions, have achieved similar speedups. In the worst case scenario, the FPGA implementation achieved either a similar or slightly better performance to the baseline (for all DL operations); where worst case scenario results from using a small dataset such as: using conjunction & disjunction on < 100 individuals, and using role & concrete (float/string) role restrictions on < 100, 000 assertions.\u0000","PeriodicalId":50914,"journal":{"name":"ACM Transactions on Embedded Computing Systems","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141116391","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TEFLON: Thermally Efficient Dataflow-Aware 3D NoC for Accelerating CNN Inferencing on Manycore PIM Architectures TEFLON:用于在多核 PIM 架构上加速 CNN 推断的热效数据流感知 3D NoC
IF 2 3区 计算机科学
ACM Transactions on Embedded Computing Systems Pub Date : 2024-05-16 DOI: 10.1145/3665279
Gaurav Narang, Chukwufumnanya Ogbogu, Jana Doppa, P. Pande
{"title":"TEFLON: Thermally Efficient Dataflow-Aware 3D NoC for Accelerating CNN Inferencing on Manycore PIM Architectures","authors":"Gaurav Narang, Chukwufumnanya Ogbogu, Jana Doppa, P. Pande","doi":"10.1145/3665279","DOIUrl":"https://doi.org/10.1145/3665279","url":null,"abstract":"\u0000 Resistive random-access memory (ReRAM) based processing-in-memory (PIM) architectures are used extensively to accelerate inferencing/training with convolutional neural networks (CNNs). Three-dimensional (3D) integration is an enabling technology to integrate many PIM cores on a single chip. In this work, we propose the design of a thermally efficient dataflow-aware monolithic 3D (M3D) NoC architecture referred to as\u0000 \u0000 TEFLON\u0000 \u0000 to accelerate CNN inferencing without creating any thermal bottlenecks.\u0000 \u0000 TEFLON\u0000 \u0000 reduces the Energy-Delay-Product (EDP) by 4\u0000 \u0000 2%\u0000 \u0000 ,\u0000 \u0000 46%\u0000 \u0000 , and 45\u0000 \u0000 %\u0000 \u0000 on an average compared to a conventional 3D mesh NoC for systems with 36-, 64-, and 100-PIM cores respectively.\u0000 \u0000 TEFLON\u0000 \u0000 reduces the peak chip temperature by 25\u0000 \u0000 K\u0000 \u0000 and improves the inference accuracy by up to 11\u0000 \u0000 %\u0000 \u0000 compared to sole performance-optimized SFC-based counterpart for inferencing with diverse deep CNN models using CIFAR-10/100 datasets on a 3D system with 100-PIM cores.\u0000","PeriodicalId":50914,"journal":{"name":"ACM Transactions on Embedded Computing Systems","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140971584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TREAFET: Temperature-Aware Real-Time Task Scheduling for FinFET based Multicores TREAFET:基于 FinFET 的多核温度感知实时任务调度
IF 2 3区 计算机科学
ACM Transactions on Embedded Computing Systems Pub Date : 2024-05-16 DOI: 10.1145/3665276
Shounak Chakraborty, Yanshul Sharma, S. Moulik
{"title":"TREAFET: Temperature-Aware Real-Time Task Scheduling for FinFET based Multicores","authors":"Shounak Chakraborty, Yanshul Sharma, S. Moulik","doi":"10.1145/3665276","DOIUrl":"https://doi.org/10.1145/3665276","url":null,"abstract":"\u0000 The recent shift in the VLSI industry from conventional MOSFET to FinFET for designing contemporary chip-multiprocessor (CMP) has noticeably improved hardware platforms’ computing capabilities, but at the cost of several thermal issues. Unlike the conventional MOSFET, FinFET devices experience a significant increase in circuit speed at a higher temperature, called temperature effect inversion (TEI), but higher temperature can also curtail the circuit lifetime due to self-heating effects (SHEs). These fundamental thermal properties of FinFET introduced a new challenge for scheduling time-critical tasks on FinFET based multicores that how to exploit TEI towards improving performance while combating SHEs. In this work,\u0000 TREAFET\u0000 , a temperature-aware real-time scheduler, attempts to exploit the TEI feature of FinFET based multicores in a time-critical computing paradigm. At first, the overall progress of individual tasks is monitored, tasks are allocated to the cores, and finally, a schedule is prepared. By considering the thermal profiles of the individual tasks and the current thermal status of the cores, hot tasks are assigned to the cold cores and vice-versa. Finally, the performance and temperature are balanced on-the-fly by incorporating a prudential voltage scaling towards exploiting TEI while guaranteeing the deadline and thermal safety. Moreover,\u0000 TREAFET\u0000 stimulates the average runtime frequency by employing an opportunistic energy-adaptive voltage spiking mechanism, in which energy saving during memory stalls at the cores is traded off during the time slice having the spiked voltage. Simulation results claim\u0000 TREAFET\u0000 maintains a safe and stable thermal status (peak temperature below 80°C) and improves frequency up to 17% over the assigned value, which ensures legitimate time-critical performance for a variety of workloads while surpassing a state-of-the-art technique. The stimulated frequency in\u0000 TREAFET\u0000 also finishes the tasks early, thus providing opportunities to save energy by power gating the cores, and achieves a 24% energy delay product (EDP) gain on average.\u0000","PeriodicalId":50914,"journal":{"name":"ACM Transactions on Embedded Computing Systems","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140969448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Coupling bit and modular arithmetic for efficient general-purpose fully homomorphic encryption 将位运算和模块运算结合起来,实现高效的通用全同态加密
IF 2 3区 计算机科学
ACM Transactions on Embedded Computing Systems Pub Date : 2024-05-16 DOI: 10.1145/3665280
E. Chielle, O. Mazonka, Homer Gamil, Michail Maniatakos
{"title":"Coupling bit and modular arithmetic for efficient general-purpose fully homomorphic encryption","authors":"E. Chielle, O. Mazonka, Homer Gamil, Michail Maniatakos","doi":"10.1145/3665280","DOIUrl":"https://doi.org/10.1145/3665280","url":null,"abstract":"Fully Homomorphic Encryption (FHE) enables computation directly on encrypted data. This property is desirable for outsourced computation of sensitive data as it relies solely on the underlying security of the cryptosystem and not in access control policies. Even though FHE is still significantly slower than unencrypted computation, practical times are possible for applications easily representable as low-order polynomials, since most FHE schemes support modular addition and multiplication over ciphertexts. If, however, an application cannot be expressed with low-order polynomials, then Boolean logic must be emulated. This bit-level arithmetic enables any computation to be performed homomorphically. Nevertheless, as it runs on top of the natively supported modular arithmetic, it has poor performance, which hinders its use in the majority of scenarios. In this work, we propose Bridging, a technique that allows conversion from bit-level to modular arithmetic and vice-versa. This enables the use of the comprehensive computation provided by bit-level arithmetic and the performance of modular arithmetic within the same application. Experimental results show that Bridging can lead to 1-2 orders of magnitude performance improvement for tested benchmarks and two real-world applications: URL denylisting and genotype imputation. Bridging performance comes from two factors: reduced number of operations and smaller multiplicative depth.","PeriodicalId":50914,"journal":{"name":"ACM Transactions on Embedded Computing Systems","volume":null,"pages":null},"PeriodicalIF":2.0,"publicationDate":"2024-05-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140968480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信