Deep Reinforcement Learning-Based Optimization Framework with Continuous Action Space for LNG Liquefaction Processes

IF 3.2 4区 工程技术 Q2 CHEMISTRY, MULTIDISCIPLINARY
Jieun Lee, Kyungtae Park
{"title":"Deep Reinforcement Learning-Based Optimization Framework with Continuous Action Space for LNG Liquefaction Processes","authors":"Jieun Lee,&nbsp;Kyungtae Park","doi":"10.1007/s11814-025-00428-x","DOIUrl":null,"url":null,"abstract":"<div><p>Recently, the application of reinforcement learning in process systems engineering has attracted significant attention recently. However, the optimization of chemical processes using this approach faces various challenges related to performance and stability. This paper presents a process optimization framework using a continuous advantage actor–critic that is modified from the existing advantage actor–critic algorithm by incorporating a normal distribution for action sampling in a continuous space. The proposed reinforcement learning-based optimization framework was found to outperform the conventional method in optimizing a single mixed refrigerant process with 10 variables, achieving a lower specific energy consumption value of 0.294 kWh/kg compared to the value of 0.307 kWh/kg obtained using the genetic algorithm. Parametric studies performed into the hyperparameters of the continuous advantage actor-critic algorithm, including the maximum episodes, learning rate, maximum action value, and structures of the neural networks, are presented to investigate their impacts on the optimization performance. The optimal specific energy consumption, namely 0.287 kWh/kg, was achieved by varying the learning rate from the base case to 0.00005. These results demonstrate that reinforcement learning can be effectively applied to the optimization of chemical processes.</p></div>","PeriodicalId":684,"journal":{"name":"Korean Journal of Chemical Engineering","volume":"42 8","pages":"1613 - 1628"},"PeriodicalIF":3.2000,"publicationDate":"2025-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Korean Journal of Chemical Engineering","FirstCategoryId":"5","ListUrlMain":"https://link.springer.com/article/10.1007/s11814-025-00428-x","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"CHEMISTRY, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

Abstract

Recently, the application of reinforcement learning in process systems engineering has attracted significant attention recently. However, the optimization of chemical processes using this approach faces various challenges related to performance and stability. This paper presents a process optimization framework using a continuous advantage actor–critic that is modified from the existing advantage actor–critic algorithm by incorporating a normal distribution for action sampling in a continuous space. The proposed reinforcement learning-based optimization framework was found to outperform the conventional method in optimizing a single mixed refrigerant process with 10 variables, achieving a lower specific energy consumption value of 0.294 kWh/kg compared to the value of 0.307 kWh/kg obtained using the genetic algorithm. Parametric studies performed into the hyperparameters of the continuous advantage actor-critic algorithm, including the maximum episodes, learning rate, maximum action value, and structures of the neural networks, are presented to investigate their impacts on the optimization performance. The optimal specific energy consumption, namely 0.287 kWh/kg, was achieved by varying the learning rate from the base case to 0.00005. These results demonstrate that reinforcement learning can be effectively applied to the optimization of chemical processes.

基于深度强化学习的LNG液化过程连续动作空间优化框架
近年来,强化学习在过程系统工程中的应用引起了人们的广泛关注。然而,使用这种方法优化化学过程面临着与性能和稳定性相关的各种挑战。本文提出了一个过程优化框架,该框架是在现有的优势行为者批评算法的基础上改进的,通过在连续空间中加入正态分布来进行动作采样。结果表明,基于强化学习的优化框架在10个变量的单一混合制冷剂过程优化中优于传统方法,比能耗值为0.294 kWh/kg,低于使用遗传算法获得的0.307 kWh/kg。对连续优势actor-critic算法的超参数进行了参数化研究,包括最大情节、学习率、最大动作值和神经网络结构,以研究它们对优化性能的影响。通过将基本情况下的学习率改变为0.00005,可以实现最优的比能耗,即0.287 kWh/kg。这些结果表明,强化学习可以有效地应用于化工过程的优化。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Korean Journal of Chemical Engineering
Korean Journal of Chemical Engineering 工程技术-工程:化工
CiteScore
4.60
自引率
11.10%
发文量
310
审稿时长
4.7 months
期刊介绍: The Korean Journal of Chemical Engineering provides a global forum for the dissemination of research in chemical engineering. The Journal publishes significant research results obtained in the Asia-Pacific region, and simultaneously introduces recent technical progress made in other areas of the world to this region. Submitted research papers must be of potential industrial significance and specifically concerned with chemical engineering. The editors will give preference to papers having a clearly stated practical scope and applicability in the areas of chemical engineering, and to those where new theoretical concepts are supported by new experimental details. The Journal also regularly publishes featured reviews on emerging and industrially important subjects of chemical engineering as well as selected papers presented at international conferences on the subjects.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信