Yuzheng Ren;Haijun Zhang;Fei Richard Yu;Wei Li;Pincan Zhao;Ying He
{"title":"具有大型语言模型的工业物联网:基于智能的强化学习方法","authors":"Yuzheng Ren;Haijun Zhang;Fei Richard Yu;Wei Li;Pincan Zhao;Ying He","doi":"10.1109/TMC.2024.3522130","DOIUrl":null,"url":null,"abstract":"Large Language Models (LLMs), as advanced AI technologies for processing and generating natural language text, bring substantial benefits to the Industrial Internet of Things (IIoT) by enhancing efficiency, decision-making, and automation. Nevertheless, their deployment faces significant obstacles due to high computational and energy demands, which often exceed the capabilities of many industrial devices. To overcome these challenges, edge-cloud collaboration has become increasingly essential, assisting in offloading LLMs tasks to reduce the computational load. However, traditional reinforcement learning (RL)-based strategies for LLMs task offloading encounter difficulties with generalization ability and defining explicit, appropriate reward functions. Therefore, in this paper, we propose a novel framework for offloading LLMs inference tasks in IIoT, utilizing a Decentralized Identifier (DID)-based identity management system for trusted task offloading. Furthermore, we introduce an intelligence-based RL (IRL) approach, which sidesteps the need for defining specific reward functions. Instead, it uses “intelligence” as a metric to evaluate cognitive improvements and adapt to varying environmental preferences, significantly improving generalizability. In our experiments, we employ the GPT-J-6B model and utilize the Human Eval dataset to assess its ability to tackle programming challenges, demonstrating the superior performance of our proposed solution compared to existing methods.","PeriodicalId":50389,"journal":{"name":"IEEE Transactions on Mobile Computing","volume":"24 5","pages":"4136-4152"},"PeriodicalIF":7.7000,"publicationDate":"2024-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Industrial Internet of Things With Large Language Models (LLMs): An Intelligence-Based Reinforcement Learning Approach\",\"authors\":\"Yuzheng Ren;Haijun Zhang;Fei Richard Yu;Wei Li;Pincan Zhao;Ying He\",\"doi\":\"10.1109/TMC.2024.3522130\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Large Language Models (LLMs), as advanced AI technologies for processing and generating natural language text, bring substantial benefits to the Industrial Internet of Things (IIoT) by enhancing efficiency, decision-making, and automation. Nevertheless, their deployment faces significant obstacles due to high computational and energy demands, which often exceed the capabilities of many industrial devices. To overcome these challenges, edge-cloud collaboration has become increasingly essential, assisting in offloading LLMs tasks to reduce the computational load. However, traditional reinforcement learning (RL)-based strategies for LLMs task offloading encounter difficulties with generalization ability and defining explicit, appropriate reward functions. Therefore, in this paper, we propose a novel framework for offloading LLMs inference tasks in IIoT, utilizing a Decentralized Identifier (DID)-based identity management system for trusted task offloading. Furthermore, we introduce an intelligence-based RL (IRL) approach, which sidesteps the need for defining specific reward functions. Instead, it uses “intelligence” as a metric to evaluate cognitive improvements and adapt to varying environmental preferences, significantly improving generalizability. In our experiments, we employ the GPT-J-6B model and utilize the Human Eval dataset to assess its ability to tackle programming challenges, demonstrating the superior performance of our proposed solution compared to existing methods.\",\"PeriodicalId\":50389,\"journal\":{\"name\":\"IEEE Transactions on Mobile Computing\",\"volume\":\"24 5\",\"pages\":\"4136-4152\"},\"PeriodicalIF\":7.7000,\"publicationDate\":\"2024-12-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Mobile Computing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10813369/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Mobile Computing","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10813369/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Industrial Internet of Things With Large Language Models (LLMs): An Intelligence-Based Reinforcement Learning Approach
Large Language Models (LLMs), as advanced AI technologies for processing and generating natural language text, bring substantial benefits to the Industrial Internet of Things (IIoT) by enhancing efficiency, decision-making, and automation. Nevertheless, their deployment faces significant obstacles due to high computational and energy demands, which often exceed the capabilities of many industrial devices. To overcome these challenges, edge-cloud collaboration has become increasingly essential, assisting in offloading LLMs tasks to reduce the computational load. However, traditional reinforcement learning (RL)-based strategies for LLMs task offloading encounter difficulties with generalization ability and defining explicit, appropriate reward functions. Therefore, in this paper, we propose a novel framework for offloading LLMs inference tasks in IIoT, utilizing a Decentralized Identifier (DID)-based identity management system for trusted task offloading. Furthermore, we introduce an intelligence-based RL (IRL) approach, which sidesteps the need for defining specific reward functions. Instead, it uses “intelligence” as a metric to evaluate cognitive improvements and adapt to varying environmental preferences, significantly improving generalizability. In our experiments, we employ the GPT-J-6B model and utilize the Human Eval dataset to assess its ability to tackle programming challenges, demonstrating the superior performance of our proposed solution compared to existing methods.
期刊介绍:
IEEE Transactions on Mobile Computing addresses key technical issues related to various aspects of mobile computing. This includes (a) architectures, (b) support services, (c) algorithm/protocol design and analysis, (d) mobile environments, (e) mobile communication systems, (f) applications, and (g) emerging technologies. Topics of interest span a wide range, covering aspects like mobile networks and hosts, mobility management, multimedia, operating system support, power management, online and mobile environments, security, scalability, reliability, and emerging technologies such as wearable computers, body area networks, and wireless sensor networks. The journal serves as a comprehensive platform for advancements in mobile computing research.