Cristiano da Costa Cunha;Wei Liu;Tim French;Ajmal Mian
{"title":"Q-Cogni: An Integrated Causal Reinforcement Learning Framework","authors":"Cristiano da Costa Cunha;Wei Liu;Tim French;Ajmal Mian","doi":"10.1109/TAI.2024.3453230","DOIUrl":null,"url":null,"abstract":"We present \n<italic>Q-Cogni</i>\n, an algorithmically integrated causal reinforcement learning framework that redesigns \n<italic>Q-Learning</i>\n to improve the learning process with causal inference. \n<italic>Q-Cogni</i>\n achieves improved policy quality and learning efficiency with a prelearned structural causal model of the environment, queried to guide the policy learning process with an understanding of cause-and-effect relationships in a state-action space. By doing so, we not only leverage the sample efficient techniques of reinforcement learning but also enable reasoning about a broader set of policies and bring higher degrees of interpretability to decisions made by the reinforcement learning agent. We apply \n<italic>Q-Cogni</i>\n on vehicle routing problem (VRP) environments including a real-world dataset of taxis in New York City using the Taxi and Limousine Commission trip record data. We show \n<italic>Q-Cogni's</i>\n capability to achieve an optimally guaranteed policy (total trip distance) in 76% of the cases when comparing to shortest-path-search methods and outperforming (shorter distances) state-of-the-art reinforcement learning algorithms in 66% of cases. Additionally, since \n<italic>Q-Cogni</i>\n does not require a complete global map, we show that it can start efficiently routing with partial information and improve as more data is collected, such as traffic disruptions and changes in destination, making it ideal for deployment in real-world dynamic settings.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"5 12","pages":"6186-6195"},"PeriodicalIF":0.0000,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on artificial intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10663687/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
We present
Q-Cogni
, an algorithmically integrated causal reinforcement learning framework that redesigns
Q-Learning
to improve the learning process with causal inference.
Q-Cogni
achieves improved policy quality and learning efficiency with a prelearned structural causal model of the environment, queried to guide the policy learning process with an understanding of cause-and-effect relationships in a state-action space. By doing so, we not only leverage the sample efficient techniques of reinforcement learning but also enable reasoning about a broader set of policies and bring higher degrees of interpretability to decisions made by the reinforcement learning agent. We apply
Q-Cogni
on vehicle routing problem (VRP) environments including a real-world dataset of taxis in New York City using the Taxi and Limousine Commission trip record data. We show
Q-Cogni's
capability to achieve an optimally guaranteed policy (total trip distance) in 76% of the cases when comparing to shortest-path-search methods and outperforming (shorter distances) state-of-the-art reinforcement learning algorithms in 66% of cases. Additionally, since
Q-Cogni
does not require a complete global map, we show that it can start efficiently routing with partial information and improve as more data is collected, such as traffic disruptions and changes in destination, making it ideal for deployment in real-world dynamic settings.