Sebastián Espinel‐Ríos, River Walser, Dongda Zhang
{"title":"鲁棒动态代谢控制的强化学习","authors":"Sebastián Espinel‐Ríos, River Walser, Dongda Zhang","doi":"10.1002/bit.70077","DOIUrl":null,"url":null,"abstract":"Dynamic metabolic control allows key metabolic fluxes to be modulated in real time, enhancing bioprocess flexibility and expanding available optimization degrees of freedom. This is achieved, for example, via targeted modulation of metabolic enzyme expression. However, identifying optimal dynamic control policies is challenging due to the generally high‐dimensional solution space and the need to manage metabolic burden and cytotoxic effects arising from inducible enzyme expression. The task is further complicated by stochastic dynamics, which reduce bioprocess reproducibility. We propose a reinforcement learning framework to derive optimal policies by allowing an agent (the controller) to interact with a surrogate dynamic model. To promote robustness, we apply domain randomization, enabling the controller to generalize across uncertainties. When transferred to an experimental system, the agent can in principle continue fine‐tuning the policy. Our framework provides an alternative to conventional model‐based control such as model predictive control, which requires model differentiation with respect to decision variables; often impractical for complex stochastic, nonlinear, stiff, and piecewise‐defined dynamics. In contrast, our approach relies on forward integration of the model, thereby simplifying the task. We demonstrate the framework in two <jats:italic>Escherichia coli</jats:italic> bioprocesses: dynamic control of acetyl‐CoA carboxylase for fatty‐acid synthesis and of adenosine triphosphatase for lactate synthesis.","PeriodicalId":9168,"journal":{"name":"Biotechnology and Bioengineering","volume":"22 1","pages":""},"PeriodicalIF":3.6000,"publicationDate":"2025-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Reinforcement Learning for Robust Dynamic Metabolic Control\",\"authors\":\"Sebastián Espinel‐Ríos, River Walser, Dongda Zhang\",\"doi\":\"10.1002/bit.70077\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Dynamic metabolic control allows key metabolic fluxes to be modulated in real time, enhancing bioprocess flexibility and expanding available optimization degrees of freedom. This is achieved, for example, via targeted modulation of metabolic enzyme expression. However, identifying optimal dynamic control policies is challenging due to the generally high‐dimensional solution space and the need to manage metabolic burden and cytotoxic effects arising from inducible enzyme expression. The task is further complicated by stochastic dynamics, which reduce bioprocess reproducibility. We propose a reinforcement learning framework to derive optimal policies by allowing an agent (the controller) to interact with a surrogate dynamic model. To promote robustness, we apply domain randomization, enabling the controller to generalize across uncertainties. When transferred to an experimental system, the agent can in principle continue fine‐tuning the policy. Our framework provides an alternative to conventional model‐based control such as model predictive control, which requires model differentiation with respect to decision variables; often impractical for complex stochastic, nonlinear, stiff, and piecewise‐defined dynamics. In contrast, our approach relies on forward integration of the model, thereby simplifying the task. We demonstrate the framework in two <jats:italic>Escherichia coli</jats:italic> bioprocesses: dynamic control of acetyl‐CoA carboxylase for fatty‐acid synthesis and of adenosine triphosphatase for lactate synthesis.\",\"PeriodicalId\":9168,\"journal\":{\"name\":\"Biotechnology and Bioengineering\",\"volume\":\"22 1\",\"pages\":\"\"},\"PeriodicalIF\":3.6000,\"publicationDate\":\"2025-10-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Biotechnology and Bioengineering\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://doi.org/10.1002/bit.70077\",\"RegionNum\":2,\"RegionCategory\":\"生物学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"BIOTECHNOLOGY & APPLIED MICROBIOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biotechnology and Bioengineering","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1002/bit.70077","RegionNum":2,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"BIOTECHNOLOGY & APPLIED MICROBIOLOGY","Score":null,"Total":0}
Reinforcement Learning for Robust Dynamic Metabolic Control
Dynamic metabolic control allows key metabolic fluxes to be modulated in real time, enhancing bioprocess flexibility and expanding available optimization degrees of freedom. This is achieved, for example, via targeted modulation of metabolic enzyme expression. However, identifying optimal dynamic control policies is challenging due to the generally high‐dimensional solution space and the need to manage metabolic burden and cytotoxic effects arising from inducible enzyme expression. The task is further complicated by stochastic dynamics, which reduce bioprocess reproducibility. We propose a reinforcement learning framework to derive optimal policies by allowing an agent (the controller) to interact with a surrogate dynamic model. To promote robustness, we apply domain randomization, enabling the controller to generalize across uncertainties. When transferred to an experimental system, the agent can in principle continue fine‐tuning the policy. Our framework provides an alternative to conventional model‐based control such as model predictive control, which requires model differentiation with respect to decision variables; often impractical for complex stochastic, nonlinear, stiff, and piecewise‐defined dynamics. In contrast, our approach relies on forward integration of the model, thereby simplifying the task. We demonstrate the framework in two Escherichia coli bioprocesses: dynamic control of acetyl‐CoA carboxylase for fatty‐acid synthesis and of adenosine triphosphatase for lactate synthesis.
期刊介绍:
Biotechnology & Bioengineering publishes Perspectives, Articles, Reviews, Mini-Reviews, and Communications to the Editor that embrace all aspects of biotechnology. These include:
-Enzyme systems and their applications, including enzyme reactors, purification, and applied aspects of protein engineering
-Animal-cell biotechnology, including media development
-Applied aspects of cellular physiology, metabolism, and energetics
-Biocatalysis and applied enzymology, including enzyme reactors, protein engineering, and nanobiotechnology
-Biothermodynamics
-Biofuels, including biomass and renewable resource engineering
-Biomaterials, including delivery systems and materials for tissue engineering
-Bioprocess engineering, including kinetics and modeling of biological systems, transport phenomena in bioreactors, bioreactor design, monitoring, and control
-Biosensors and instrumentation
-Computational and systems biology, including bioinformatics and genomic/proteomic studies
-Environmental biotechnology, including biofilms, algal systems, and bioremediation
-Metabolic and cellular engineering
-Plant-cell biotechnology
-Spectroscopic and other analytical techniques for biotechnological applications
-Synthetic biology
-Tissue engineering, stem-cell bioengineering, regenerative medicine, gene therapy and delivery systems
The editors will consider papers for publication based on novelty, their immediate or future impact on biotechnological processes, and their contribution to the advancement of biochemical engineering science. Submission of papers dealing with routine aspects of bioprocessing, description of established equipment, and routine applications of established methodologies (e.g., control strategies, modeling, experimental methods) is discouraged. Theoretical papers will be judged based on the novelty of the approach and their potential impact, or on their novel capability to predict and elucidate experimental observations.