Nicholas T Williams, Katherine L Hoffman, Iván Díaz, Kara E Rudolph
{"title":"从纵向数据中学习最佳动态治疗方案。","authors":"Nicholas T Williams, Katherine L Hoffman, Iván Díaz, Kara E Rudolph","doi":"10.1093/aje/kwae122","DOIUrl":null,"url":null,"abstract":"<p><p>Investigators often report estimates of the average treatment effect (ATE). While the ATE summarizes the effect of a treatment on average, it does not provide any information about the effect of treatment within any individual. A treatment strategy that uses an individual's information to tailor treatment to maximize benefit is known as an optimal dynamic treatment rule (ODTR). Treatment, however, is typically not limited to a single point in time; consequently, learning an optimal rule for a time-varying treatment may involve not just learning the extent to which the comparative treatments' benefits vary across the characteristics of individuals, but also learning the extent to which the comparative treatments' benefits vary as relevant circumstances evolve within an individual. The goal of this paper is to provide a tutorial for estimating ODTR from longitudinal observational and clinical trial data for applied researchers. We describe an approach that uses a doubly robust unbiased transformation of the conditional ATE. We then learn a time-varying ODTR for when to increase buprenorphine-naloxone dose to minimize a return to regular opioid use among patients with opioid use disorder. Our analysis highlights the utility of ODTRs in the context of sequential decision-making: The learned ODTR outperforms a clinically defined strategy. This article is part of a Special Collection on Pharmacoepidemiology.</p>","PeriodicalId":7472,"journal":{"name":"American journal of epidemiology","volume":" ","pages":"1768-1775"},"PeriodicalIF":5.0000,"publicationDate":"2024-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11637529/pdf/","citationCount":"0","resultStr":"{\"title\":\"Learning optimal dynamic treatment regimes from longitudinal data.\",\"authors\":\"Nicholas T Williams, Katherine L Hoffman, Iván Díaz, Kara E Rudolph\",\"doi\":\"10.1093/aje/kwae122\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Investigators often report estimates of the average treatment effect (ATE). While the ATE summarizes the effect of a treatment on average, it does not provide any information about the effect of treatment within any individual. A treatment strategy that uses an individual's information to tailor treatment to maximize benefit is known as an optimal dynamic treatment rule (ODTR). Treatment, however, is typically not limited to a single point in time; consequently, learning an optimal rule for a time-varying treatment may involve not just learning the extent to which the comparative treatments' benefits vary across the characteristics of individuals, but also learning the extent to which the comparative treatments' benefits vary as relevant circumstances evolve within an individual. The goal of this paper is to provide a tutorial for estimating ODTR from longitudinal observational and clinical trial data for applied researchers. We describe an approach that uses a doubly robust unbiased transformation of the conditional ATE. We then learn a time-varying ODTR for when to increase buprenorphine-naloxone dose to minimize a return to regular opioid use among patients with opioid use disorder. Our analysis highlights the utility of ODTRs in the context of sequential decision-making: The learned ODTR outperforms a clinically defined strategy. This article is part of a Special Collection on Pharmacoepidemiology.</p>\",\"PeriodicalId\":7472,\"journal\":{\"name\":\"American journal of epidemiology\",\"volume\":\" \",\"pages\":\"1768-1775\"},\"PeriodicalIF\":5.0000,\"publicationDate\":\"2024-12-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11637529/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"American journal of epidemiology\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1093/aje/kwae122\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"PUBLIC, ENVIRONMENTAL & OCCUPATIONAL HEALTH\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"American journal of epidemiology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1093/aje/kwae122","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PUBLIC, ENVIRONMENTAL & OCCUPATIONAL HEALTH","Score":null,"Total":0}
引用次数: 0
摘要
研究通常报告平均治疗效果(ATE)的估计值。虽然 ATE 总结了平均治疗效果,但它并没有提供任何关于个体治疗效果的信息。利用个体信息来调整治疗以实现收益最大化的治疗策略被称为最佳动态治疗规则(ODTR)。然而,治疗通常并不局限于一个单一的时间点;因此,学习时变治疗的最优规则可能不仅涉及学习比较治疗在不同个体特征下的收益变化程度,还涉及学习比较治疗在个体内部相关情况发生变化时的收益变化程度。本文旨在为应用研究人员提供从纵向观察和临床试验数据中估算 ODTR 的教程。我们介绍了一种使用双重稳健无偏变换条件平均治疗效果的方法。然后,我们学习了一种随时间变化的 ODTR,即何时增加丁丙诺啡-纳洛酮(BUP-NX)的剂量,以尽量减少阿片类药物使用障碍患者恢复正常阿片类药物使用。我们的分析凸显了 ODTR 在顺序决策中的实用性:学习到的 ODTR 优于临床定义的策略。
Learning optimal dynamic treatment regimes from longitudinal data.
Investigators often report estimates of the average treatment effect (ATE). While the ATE summarizes the effect of a treatment on average, it does not provide any information about the effect of treatment within any individual. A treatment strategy that uses an individual's information to tailor treatment to maximize benefit is known as an optimal dynamic treatment rule (ODTR). Treatment, however, is typically not limited to a single point in time; consequently, learning an optimal rule for a time-varying treatment may involve not just learning the extent to which the comparative treatments' benefits vary across the characteristics of individuals, but also learning the extent to which the comparative treatments' benefits vary as relevant circumstances evolve within an individual. The goal of this paper is to provide a tutorial for estimating ODTR from longitudinal observational and clinical trial data for applied researchers. We describe an approach that uses a doubly robust unbiased transformation of the conditional ATE. We then learn a time-varying ODTR for when to increase buprenorphine-naloxone dose to minimize a return to regular opioid use among patients with opioid use disorder. Our analysis highlights the utility of ODTRs in the context of sequential decision-making: The learned ODTR outperforms a clinically defined strategy. This article is part of a Special Collection on Pharmacoepidemiology.
期刊介绍:
The American Journal of Epidemiology is the oldest and one of the premier epidemiologic journals devoted to the publication of empirical research findings, opinion pieces, and methodological developments in the field of epidemiologic research.
It is a peer-reviewed journal aimed at both fellow epidemiologists and those who use epidemiologic data, including public health workers and clinicians.