{"title":"FedTinyWolf - 一种内存高效的联合嵌入式学习机制","authors":"Subhrangshu Adhikary;Subhayu Dutta","doi":"10.1109/LES.2024.3462638","DOIUrl":null,"url":null,"abstract":"Embedded intelligence is a challenging field in engineering given its resource-constrained environment which regular machine learning algorithms demand. Most embedded intelligence models are trained on a computer and then the learned parameters are transferred to the embedded devices to enable decision making. Although training the model within a microcontroller is possible, the state-of-the-art method requires further optimization. Moreover, federated learning (FL) is used in the state of the art to protect data privacy while training a deep learning model at edge level. Embedded learning models require memory enhancements to improve on-device FL. In this experiment, we have performed memory enhancement of gray wolf optimizer after finding it suitable for the purpose and implemented it to create edge-level, resource-efficient, data privacy preserving on-device federated training of embedded intelligence models. The performances are benchmarked on 13 open-sourced datasets showing a mean 10.8% accuracy enhancement.","PeriodicalId":56143,"journal":{"name":"IEEE Embedded Systems Letters","volume":"16 4","pages":"513-516"},"PeriodicalIF":1.7000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"FedTinyWolf—A Memory Efficient Federated Embedded Learning Mechanism\",\"authors\":\"Subhrangshu Adhikary;Subhayu Dutta\",\"doi\":\"10.1109/LES.2024.3462638\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Embedded intelligence is a challenging field in engineering given its resource-constrained environment which regular machine learning algorithms demand. Most embedded intelligence models are trained on a computer and then the learned parameters are transferred to the embedded devices to enable decision making. Although training the model within a microcontroller is possible, the state-of-the-art method requires further optimization. Moreover, federated learning (FL) is used in the state of the art to protect data privacy while training a deep learning model at edge level. Embedded learning models require memory enhancements to improve on-device FL. In this experiment, we have performed memory enhancement of gray wolf optimizer after finding it suitable for the purpose and implemented it to create edge-level, resource-efficient, data privacy preserving on-device federated training of embedded intelligence models. The performances are benchmarked on 13 open-sourced datasets showing a mean 10.8% accuracy enhancement.\",\"PeriodicalId\":56143,\"journal\":{\"name\":\"IEEE Embedded Systems Letters\",\"volume\":\"16 4\",\"pages\":\"513-516\"},\"PeriodicalIF\":1.7000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Embedded Systems Letters\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10681455/\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Embedded Systems Letters","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10681455/","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
Embedded intelligence is a challenging field in engineering given its resource-constrained environment which regular machine learning algorithms demand. Most embedded intelligence models are trained on a computer and then the learned parameters are transferred to the embedded devices to enable decision making. Although training the model within a microcontroller is possible, the state-of-the-art method requires further optimization. Moreover, federated learning (FL) is used in the state of the art to protect data privacy while training a deep learning model at edge level. Embedded learning models require memory enhancements to improve on-device FL. In this experiment, we have performed memory enhancement of gray wolf optimizer after finding it suitable for the purpose and implemented it to create edge-level, resource-efficient, data privacy preserving on-device federated training of embedded intelligence models. The performances are benchmarked on 13 open-sourced datasets showing a mean 10.8% accuracy enhancement.
期刊介绍:
The IEEE Embedded Systems Letters (ESL), provides a forum for rapid dissemination of latest technical advances in embedded systems and related areas in embedded software. The emphasis is on models, methods, and tools that ensure secure, correct, efficient and robust design of embedded systems and their applications.