{"title":"Process Control in Semiconductor Manufacturing Based on Deep Distributional Soft Actor-Critic Reinforcement Learning","authors":"Bangxu Liu;Dewen Zhao;Xinchun Lu;Yuhong Liu","doi":"10.1109/TSM.2025.3539223","DOIUrl":null,"url":null,"abstract":"The quality of semiconductor fabrication processes is typically degraded by variations in the manufacturing environment, which can be suppressed by run-to-run (R2R) control schemes. The performance of controlling systems to the produce process which is always highly complex and nonlinear physical model thus is strongly associated with the controlling strategy. However, previous works focusing on less complex semiconductor fabrication processes or linear controlling strategy are both hard to extend the application scenario. A novel structure for a R2R control system based on a distributed form of deep reinforcement learning (DRL), namely, distributional soft actor-critic (DSAC) DRL with twin-value distribution learning, is proposed for multizone pressure control in the chemical mechanical planarization (CMP) process, which is one of the most crucial manufacturing processes for the fabrication of ultra-large integrated circuits (ICs). In addition, several optimization algorithms for DRL, such as twin value distribution learning, are applied, further improving DSAC DRL to enhance the control performance. Compared with other reinforcement learning (RL)-based controllers, the proposed RL control policy achieves better control performance when tested using a multizone CMP virtual metrology (VM) model based on long short-term memory (LSTM) and one-dimensional convolutional neural network (1DCNN) architectures. This deep neural network (DNN) VM model, which is applied for the first time here to test the proposed DRL-based R2R controller for semiconductor manufacturing, is designed to preserve the complexity and nonlinearity of the CMP process by using data recorded from a practical manufacturing process at an IC fabrication facility in Tianjin, China. The novel model-free controlling schemes combined with the new VM model can be used in different R2R application scenarios. Meanwhile the results achieved using the proposed DRL control strategy strongly support its potential application in modern industrial semiconductor manufacturing and offer practical guidance for the further development of CMP procedures.","PeriodicalId":451,"journal":{"name":"IEEE Transactions on Semiconductor Manufacturing","volume":"38 2","pages":"210-231"},"PeriodicalIF":2.3000,"publicationDate":"2025-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Semiconductor Manufacturing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10878415/","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
The quality of semiconductor fabrication processes is typically degraded by variations in the manufacturing environment, which can be suppressed by run-to-run (R2R) control schemes. The performance of controlling systems to the produce process which is always highly complex and nonlinear physical model thus is strongly associated with the controlling strategy. However, previous works focusing on less complex semiconductor fabrication processes or linear controlling strategy are both hard to extend the application scenario. A novel structure for a R2R control system based on a distributed form of deep reinforcement learning (DRL), namely, distributional soft actor-critic (DSAC) DRL with twin-value distribution learning, is proposed for multizone pressure control in the chemical mechanical planarization (CMP) process, which is one of the most crucial manufacturing processes for the fabrication of ultra-large integrated circuits (ICs). In addition, several optimization algorithms for DRL, such as twin value distribution learning, are applied, further improving DSAC DRL to enhance the control performance. Compared with other reinforcement learning (RL)-based controllers, the proposed RL control policy achieves better control performance when tested using a multizone CMP virtual metrology (VM) model based on long short-term memory (LSTM) and one-dimensional convolutional neural network (1DCNN) architectures. This deep neural network (DNN) VM model, which is applied for the first time here to test the proposed DRL-based R2R controller for semiconductor manufacturing, is designed to preserve the complexity and nonlinearity of the CMP process by using data recorded from a practical manufacturing process at an IC fabrication facility in Tianjin, China. The novel model-free controlling schemes combined with the new VM model can be used in different R2R application scenarios. Meanwhile the results achieved using the proposed DRL control strategy strongly support its potential application in modern industrial semiconductor manufacturing and offer practical guidance for the further development of CMP procedures.
期刊介绍:
The IEEE Transactions on Semiconductor Manufacturing addresses the challenging problems of manufacturing complex microelectronic components, especially very large scale integrated circuits (VLSI). Manufacturing these products requires precision micropatterning, precise control of materials properties, ultraclean work environments, and complex interactions of chemical, physical, electrical and mechanical processes.