{"title":"Discovery of Customized Dispatching Rule for Single-Machine Production Scheduling Using Deep Reinforcement Learning","authors":"P. C. Chua, S. K. Moon, Y. Ng, H. Ng, Manel Lopez","doi":"10.1115/detc2022-89829","DOIUrl":null,"url":null,"abstract":"\n A dispatching rule has become one of the most widely used approaches in producing scheduling due to its low time complexities and the ability to respond to dynamic changes in production. However, there is no one dispatching rule that dominates the others for the performance measure of interest. By modelling the selection of a dispatching rule to transit from one production state to another using a Markov decision process, current methods involving reinforcement learning make use of a predefined list of dispatching rules, which may limit the optimization of a specified performance measure. Greater flexibility can be achieved by creating customized dispatching rules through the important selection of production parameters for the performance measure in question. Using parameters obtained readily within the digital twin setting, this paper investigates the application of deep reinforcement learning to select customized dispatching rules formed by weighted combinations of production parameters on a single machine production scheduling problem. Due to the curse of dimensionality of storing Q values for all possible production states in a Q-table, a deep Q network is trained for the dynamic selection of the customized dispatching rules. Preliminary results show its effectiveness in minimizing total tardiness and outperform well-known existing dispatching rules.","PeriodicalId":382970,"journal":{"name":"Volume 2: 42nd Computers and Information in Engineering Conference (CIE)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Volume 2: 42nd Computers and Information in Engineering Conference (CIE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1115/detc2022-89829","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
A dispatching rule has become one of the most widely used approaches in producing scheduling due to its low time complexities and the ability to respond to dynamic changes in production. However, there is no one dispatching rule that dominates the others for the performance measure of interest. By modelling the selection of a dispatching rule to transit from one production state to another using a Markov decision process, current methods involving reinforcement learning make use of a predefined list of dispatching rules, which may limit the optimization of a specified performance measure. Greater flexibility can be achieved by creating customized dispatching rules through the important selection of production parameters for the performance measure in question. Using parameters obtained readily within the digital twin setting, this paper investigates the application of deep reinforcement learning to select customized dispatching rules formed by weighted combinations of production parameters on a single machine production scheduling problem. Due to the curse of dimensionality of storing Q values for all possible production states in a Q-table, a deep Q network is trained for the dynamic selection of the customized dispatching rules. Preliminary results show its effectiveness in minimizing total tardiness and outperform well-known existing dispatching rules.