Correction to deep reinforcement learning-based ordering mechanism for performance optimization in multi-echelon supply chains

IF 1.3 4区 数学 Q3 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS
Dony S. Kurian, V. Madhusudanan Pillai
{"title":"Correction to deep reinforcement learning-based ordering mechanism for performance optimization in multi-echelon supply chains","authors":"Dony S. Kurian,&nbsp;V. Madhusudanan Pillai","doi":"10.1002/asmb.2838","DOIUrl":null,"url":null,"abstract":"<p>This paper addresses and acknowledges the valuable feedback provided by Dr. Deniz Preil in response to the recent study conducted by Kurian et al which investigates the application of proximal policy optimization (PPO) to determine dynamic ordering policies within multi-echelon supply chains. The first comment raised by Dr. Preil motivated an examination of the training and evaluation procedures in Experiments 2, 3, and 4. The Experiments 2 and 3 were reworked to address this, allowing the seed to vary for every training iteration, resulting in refined outcomes while there was no need of reworking of Experiment 4. The second comment focused on the benchmarking strategies involving the 1-1 policy and the order-up-to (OUT) policy, clarifying the distinctions between the two policies and justifying the use of the 1-1 policy for benchmarking in Experiment 4. The implementation of the widely accepted OUT policy was explained, highlighting the meaningful rationale behind its use. These discussions aim to enhance the methodology employed by Kurian et al and strengthen the implications of the findings within the domain of supply chain ordering management.</p>","PeriodicalId":55495,"journal":{"name":"Applied Stochastic Models in Business and Industry","volume":null,"pages":null},"PeriodicalIF":1.3000,"publicationDate":"2023-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Stochastic Models in Business and Industry","FirstCategoryId":"100","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/asmb.2838","RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"MATHEMATICS, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0

Abstract

This paper addresses and acknowledges the valuable feedback provided by Dr. Deniz Preil in response to the recent study conducted by Kurian et al which investigates the application of proximal policy optimization (PPO) to determine dynamic ordering policies within multi-echelon supply chains. The first comment raised by Dr. Preil motivated an examination of the training and evaluation procedures in Experiments 2, 3, and 4. The Experiments 2 and 3 were reworked to address this, allowing the seed to vary for every training iteration, resulting in refined outcomes while there was no need of reworking of Experiment 4. The second comment focused on the benchmarking strategies involving the 1-1 policy and the order-up-to (OUT) policy, clarifying the distinctions between the two policies and justifying the use of the 1-1 policy for benchmarking in Experiment 4. The implementation of the widely accepted OUT policy was explained, highlighting the meaningful rationale behind its use. These discussions aim to enhance the methodology employed by Kurian et al and strengthen the implications of the findings within the domain of supply chain ordering management.

修正基于深度强化学习的订购机制,以优化多供应链的性能
本文讨论并感谢 Deniz Preil 博士针对 Kurian 等人最近开展的研究提供的宝贵反馈意见,该研究调查了近端策略优化 (PPO) 在多十轴供应链中用于确定动态订购策略的应用。Preil 博士提出的第一条意见促使我们对实验 2、3 和 4 中的培训和评估程序进行检查。为了解决这个问题,对实验 2 和实验 3 进行了重新设计,允许每次训练迭代的种子都不同,结果更加完善,而实验 4 则无需重新设计。第二条评论主要针对涉及 1-1 策略和阶次提升(OUT)策略的基准测试策略,阐明了这两种策略之间的区别,并证明了在实验 4 中使用 1-1 策略进行基准测试的合理性。此外,还对广为接受的 OUT 政策的实施进行了解释,强调了其使用背后的意义。这些讨论旨在改进 Kurian 等人采用的方法,并加强研究结果在供应链订货管理领域的影响。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
2.70
自引率
0.00%
发文量
67
审稿时长
>12 weeks
期刊介绍: ASMBI - Applied Stochastic Models in Business and Industry (formerly Applied Stochastic Models and Data Analysis) was first published in 1985, publishing contributions in the interface between stochastic modelling, data analysis and their applications in business, finance, insurance, management and production. In 2007 ASMBI became the official journal of the International Society for Business and Industrial Statistics (www.isbis.org). The main objective is to publish papers, both technical and practical, presenting new results which solve real-life problems or have great potential in doing so. Mathematical rigour, innovative stochastic modelling and sound applications are the key ingredients of papers to be published, after a very selective review process. The journal is very open to new ideas, like Data Science and Big Data stemming from problems in business and industry or uncertainty quantification in engineering, as well as more traditional ones, like reliability, quality control, design of experiments, managerial processes, supply chains and inventories, insurance, econometrics, financial modelling (provided the papers are related to real problems). The journal is interested also in papers addressing the effects of business and industrial decisions on the environment, healthcare, social life. State-of-the art computational methods are very welcome as well, when combined with sound applications and innovative models.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信