Illhoe Hwang, H. Cho, S. Hong, Junhui Lee, SeokJoong Kim, Y. Jang
{"title":"基于q学习的OHT系统路径引导与车辆分配","authors":"Illhoe Hwang, H. Cho, S. Hong, Junhui Lee, SeokJoong Kim, Y. Jang","doi":"10.1109/ASMC49169.2020.9185357","DOIUrl":null,"url":null,"abstract":"We present a reinforcement learning-based algorithm for route guidance and vehicle assignment of an overhead hoist transport system, a typical form of automated material handling system in semiconductor fabrication facilities (fabs). As the size of the fab increases, so does the number of vehicles required to operate in the fab. The algorithm most commonly used in industry, a mathematical optimization-based algorithm that constantly seeks the shortest routes, has been proven ineffective in dealing with fabs operating around 1,000 vehicles or more. In this paper, we introduce Q-learning, a reinforcement learning-based algorithm for route guidance and vehicle assignment. Q-learning dynamically reroutes the vehicles based on the congestion and traffic conditions. It also assigns vehicles to tasks based on the overall congestion of the track. We show that the proposed algorithm is considerably more effective than the existing algorithm in an actual fab-scale experiment. Moreover, we illustrate that the Q-learning-based algorithm is more effective in designing the track layouts.","PeriodicalId":6771,"journal":{"name":"2020 31st Annual SEMI Advanced Semiconductor Manufacturing Conference (ASMC)","volume":"31 1","pages":"1-6"},"PeriodicalIF":0.0000,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Q-learning-based route-guidance and vehicle assignment for OHT systems in semiconductor fabs\",\"authors\":\"Illhoe Hwang, H. Cho, S. Hong, Junhui Lee, SeokJoong Kim, Y. Jang\",\"doi\":\"10.1109/ASMC49169.2020.9185357\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We present a reinforcement learning-based algorithm for route guidance and vehicle assignment of an overhead hoist transport system, a typical form of automated material handling system in semiconductor fabrication facilities (fabs). As the size of the fab increases, so does the number of vehicles required to operate in the fab. The algorithm most commonly used in industry, a mathematical optimization-based algorithm that constantly seeks the shortest routes, has been proven ineffective in dealing with fabs operating around 1,000 vehicles or more. In this paper, we introduce Q-learning, a reinforcement learning-based algorithm for route guidance and vehicle assignment. Q-learning dynamically reroutes the vehicles based on the congestion and traffic conditions. It also assigns vehicles to tasks based on the overall congestion of the track. We show that the proposed algorithm is considerably more effective than the existing algorithm in an actual fab-scale experiment. Moreover, we illustrate that the Q-learning-based algorithm is more effective in designing the track layouts.\",\"PeriodicalId\":6771,\"journal\":{\"name\":\"2020 31st Annual SEMI Advanced Semiconductor Manufacturing Conference (ASMC)\",\"volume\":\"31 1\",\"pages\":\"1-6\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 31st Annual SEMI Advanced Semiconductor Manufacturing Conference (ASMC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ASMC49169.2020.9185357\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 31st Annual SEMI Advanced Semiconductor Manufacturing Conference (ASMC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ASMC49169.2020.9185357","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Q-learning-based route-guidance and vehicle assignment for OHT systems in semiconductor fabs
We present a reinforcement learning-based algorithm for route guidance and vehicle assignment of an overhead hoist transport system, a typical form of automated material handling system in semiconductor fabrication facilities (fabs). As the size of the fab increases, so does the number of vehicles required to operate in the fab. The algorithm most commonly used in industry, a mathematical optimization-based algorithm that constantly seeks the shortest routes, has been proven ineffective in dealing with fabs operating around 1,000 vehicles or more. In this paper, we introduce Q-learning, a reinforcement learning-based algorithm for route guidance and vehicle assignment. Q-learning dynamically reroutes the vehicles based on the congestion and traffic conditions. It also assigns vehicles to tasks based on the overall congestion of the track. We show that the proposed algorithm is considerably more effective than the existing algorithm in an actual fab-scale experiment. Moreover, we illustrate that the Q-learning-based algorithm is more effective in designing the track layouts.