{"title":"DiReDi:面向 AIoT 应用的蒸馏和反向蒸馏","authors":"Chen Sun, Qing Tong, Wenshuang Yang, Wenqi Zhang","doi":"arxiv-2409.08308","DOIUrl":null,"url":null,"abstract":"Typically, the significant efficiency can be achieved by deploying different\nedge AI models in various real world scenarios while a few large models manage\nthose edge AI models remotely from cloud servers. However, customizing edge AI\nmodels for each user's specific application or extending current models to new\napplication scenarios remains a challenge. Inappropriate local training or fine\ntuning of edge AI models by users can lead to model malfunction, potentially\nresulting in legal issues for the manufacturer. To address aforementioned\nissues, this paper proposes an innovative framework called \"DiReD\", which\ninvolves knowledge DIstillation & REverse DIstillation. In the initial step, an\nedge AI model is trained with presumed data and a KD process using the cloud AI\nmodel in the upper management cloud server. This edge AI model is then\ndispatched to edge AI devices solely for inference in the user's application\nscenario. When the user needs to update the edge AI model to better fit the\nactual scenario, the reverse distillation (RD) process is employed to extract\nthe knowledge: the difference between user preferences and the manufacturer's\npresumptions from the edge AI model using the user's exclusive data. Only the\nextracted knowledge is reported back to the upper management cloud server to\nupdate the cloud AI model, thus protecting user privacy by not using any\nexclusive data. The updated cloud AI can then update the edge AI model with the\nextended knowledge. Simulation results demonstrate that the proposed \"DiReDi\"\nframework allows the manufacturer to update the user model by learning new\nknowledge from the user's actual scenario with private data. The initial\nredundant knowledge is reduced since the retraining emphasizes user private\ndata.","PeriodicalId":501422,"journal":{"name":"arXiv - CS - Distributed, Parallel, and Cluster Computing","volume":"64 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"DiReDi: Distillation and Reverse Distillation for AIoT Applications\",\"authors\":\"Chen Sun, Qing Tong, Wenshuang Yang, Wenqi Zhang\",\"doi\":\"arxiv-2409.08308\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Typically, the significant efficiency can be achieved by deploying different\\nedge AI models in various real world scenarios while a few large models manage\\nthose edge AI models remotely from cloud servers. However, customizing edge AI\\nmodels for each user's specific application or extending current models to new\\napplication scenarios remains a challenge. Inappropriate local training or fine\\ntuning of edge AI models by users can lead to model malfunction, potentially\\nresulting in legal issues for the manufacturer. To address aforementioned\\nissues, this paper proposes an innovative framework called \\\"DiReD\\\", which\\ninvolves knowledge DIstillation & REverse DIstillation. In the initial step, an\\nedge AI model is trained with presumed data and a KD process using the cloud AI\\nmodel in the upper management cloud server. This edge AI model is then\\ndispatched to edge AI devices solely for inference in the user's application\\nscenario. When the user needs to update the edge AI model to better fit the\\nactual scenario, the reverse distillation (RD) process is employed to extract\\nthe knowledge: the difference between user preferences and the manufacturer's\\npresumptions from the edge AI model using the user's exclusive data. Only the\\nextracted knowledge is reported back to the upper management cloud server to\\nupdate the cloud AI model, thus protecting user privacy by not using any\\nexclusive data. The updated cloud AI can then update the edge AI model with the\\nextended knowledge. Simulation results demonstrate that the proposed \\\"DiReDi\\\"\\nframework allows the manufacturer to update the user model by learning new\\nknowledge from the user's actual scenario with private data. The initial\\nredundant knowledge is reduced since the retraining emphasizes user private\\ndata.\",\"PeriodicalId\":501422,\"journal\":{\"name\":\"arXiv - CS - Distributed, Parallel, and Cluster Computing\",\"volume\":\"64 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Distributed, Parallel, and Cluster Computing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.08308\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Distributed, Parallel, and Cluster Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.08308","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
DiReDi: Distillation and Reverse Distillation for AIoT Applications
Typically, the significant efficiency can be achieved by deploying different
edge AI models in various real world scenarios while a few large models manage
those edge AI models remotely from cloud servers. However, customizing edge AI
models for each user's specific application or extending current models to new
application scenarios remains a challenge. Inappropriate local training or fine
tuning of edge AI models by users can lead to model malfunction, potentially
resulting in legal issues for the manufacturer. To address aforementioned
issues, this paper proposes an innovative framework called "DiReD", which
involves knowledge DIstillation & REverse DIstillation. In the initial step, an
edge AI model is trained with presumed data and a KD process using the cloud AI
model in the upper management cloud server. This edge AI model is then
dispatched to edge AI devices solely for inference in the user's application
scenario. When the user needs to update the edge AI model to better fit the
actual scenario, the reverse distillation (RD) process is employed to extract
the knowledge: the difference between user preferences and the manufacturer's
presumptions from the edge AI model using the user's exclusive data. Only the
extracted knowledge is reported back to the upper management cloud server to
update the cloud AI model, thus protecting user privacy by not using any
exclusive data. The updated cloud AI can then update the edge AI model with the
extended knowledge. Simulation results demonstrate that the proposed "DiReDi"
framework allows the manufacturer to update the user model by learning new
knowledge from the user's actual scenario with private data. The initial
redundant knowledge is reduced since the retraining emphasizes user private
data.