在线抑郁症筛查后自动结果反馈效果的异质性:DISCOVER试验的二次机器学习分析。

IF 2
JMIR AI Pub Date : 2025-08-21 DOI:10.2196/70001
Matthias Klee, Byron C Jaeger, Franziska Sikorski, Bernd Löwe, Sebastian Kohlmann
{"title":"在线抑郁症筛查后自动结果反馈效果的异质性:DISCOVER试验的二次机器学习分析。","authors":"Matthias Klee, Byron C Jaeger, Franziska Sikorski, Bernd Löwe, Sebastian Kohlmann","doi":"10.2196/70001","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Online depression screening tools may increase uptake of evidence-based care and consequently lead to symptom reduction. However, results of the DISCOVER trial suggested no effect of automated results feedback compared with no feedback after online depression screening on depressive symptom reduction six months after screening. Interpersonal variation in symptom representation, health care needs, and treatment preferences may nonetheless have led to differential response to feedback mode on an individual level.</p><p><strong>Objective: </strong>The aim of this study was to examine heterogeneity of treatment effects (HTE), that is, differential responses to two feedback modes (tailored or nontailored) versus no feedback (control) following online depression screening.</p><p><strong>Methods: </strong>We used causal forests, a machine learning method that applies recursive partitioning to estimate conditional average treatment effects (CATEs). In this secondary data analysis of the DISCOVER trial, eligible participants screened positive for at least moderate depression severity but had not been diagnosed or treated for depression in the preceding year. The primary outcome was heterogeneity in depression severity change, over a and six-month follow up period, measured with the Patient Health Questionnaire-9. Analysis comprised exploration of average treatment effects (ATE), HTE, operationalized with the area under the targeting operator characteristic curve (AUTOC), and differences in ATE when allocating feedback based on predicted CATE. We extracted top predictors of depression severity change, given feedback and explored high-CATE covariate profiles. Prior to analysis, data was split into training and test sets (1:1) to minimize the risk of overfitting and evaluate predictions in held-out test data.</p><p><strong>Results: </strong>Data from 946 participants of the DISCOVER trial without missing data were analyzed. We did not detect HTE; no versus nontailored feedback, AUTOC -0.48 (95% CI -1.62 to 0.67, P=.41); no versus tailored feedback, AUTOC 0.06 (95% CI -1.21 to 1.33, P=.93); and no versus any feedback, AUTOC -0.20 (95% CI -1.30 to 0.89, P=.72). There was no evidence of alteration to the ATE in the test set when allocating feedback (tailored or nontailored) based on the predicted CATE. By examining covariate profiles, we observed a potentially detrimental role of control beliefs, given feedback compared with no feedback.</p><p><strong>Conclusions: </strong>We applied causal forests to describe higher-level interactions among a broad range of predictors to detect HTE. In absence of evidence for HTE, treatment prioritization based on trained models did not improve ATEs. We did not find evidence of harm or benefit from providing tailored or nontailored feedback after online depression screening regarding depression severity change after six months. Future studies may test whether screening alone prompts behavioral activation and downstream depression severity reduction, considering the observed uniform changes across groups.</p>","PeriodicalId":73551,"journal":{"name":"JMIR AI","volume":"4 ","pages":"e70001"},"PeriodicalIF":2.0000,"publicationDate":"2025-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12375799/pdf/","citationCount":"0","resultStr":"{\"title\":\"Heterogeneity in Effects of Automated Results Feedback After Online Depression Screening: Secondary Machine-Learning Based Analysis of the DISCOVER Trial.\",\"authors\":\"Matthias Klee, Byron C Jaeger, Franziska Sikorski, Bernd Löwe, Sebastian Kohlmann\",\"doi\":\"10.2196/70001\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Background: </strong>Online depression screening tools may increase uptake of evidence-based care and consequently lead to symptom reduction. However, results of the DISCOVER trial suggested no effect of automated results feedback compared with no feedback after online depression screening on depressive symptom reduction six months after screening. Interpersonal variation in symptom representation, health care needs, and treatment preferences may nonetheless have led to differential response to feedback mode on an individual level.</p><p><strong>Objective: </strong>The aim of this study was to examine heterogeneity of treatment effects (HTE), that is, differential responses to two feedback modes (tailored or nontailored) versus no feedback (control) following online depression screening.</p><p><strong>Methods: </strong>We used causal forests, a machine learning method that applies recursive partitioning to estimate conditional average treatment effects (CATEs). In this secondary data analysis of the DISCOVER trial, eligible participants screened positive for at least moderate depression severity but had not been diagnosed or treated for depression in the preceding year. The primary outcome was heterogeneity in depression severity change, over a and six-month follow up period, measured with the Patient Health Questionnaire-9. Analysis comprised exploration of average treatment effects (ATE), HTE, operationalized with the area under the targeting operator characteristic curve (AUTOC), and differences in ATE when allocating feedback based on predicted CATE. We extracted top predictors of depression severity change, given feedback and explored high-CATE covariate profiles. Prior to analysis, data was split into training and test sets (1:1) to minimize the risk of overfitting and evaluate predictions in held-out test data.</p><p><strong>Results: </strong>Data from 946 participants of the DISCOVER trial without missing data were analyzed. We did not detect HTE; no versus nontailored feedback, AUTOC -0.48 (95% CI -1.62 to 0.67, P=.41); no versus tailored feedback, AUTOC 0.06 (95% CI -1.21 to 1.33, P=.93); and no versus any feedback, AUTOC -0.20 (95% CI -1.30 to 0.89, P=.72). There was no evidence of alteration to the ATE in the test set when allocating feedback (tailored or nontailored) based on the predicted CATE. By examining covariate profiles, we observed a potentially detrimental role of control beliefs, given feedback compared with no feedback.</p><p><strong>Conclusions: </strong>We applied causal forests to describe higher-level interactions among a broad range of predictors to detect HTE. In absence of evidence for HTE, treatment prioritization based on trained models did not improve ATEs. We did not find evidence of harm or benefit from providing tailored or nontailored feedback after online depression screening regarding depression severity change after six months. Future studies may test whether screening alone prompts behavioral activation and downstream depression severity reduction, considering the observed uniform changes across groups.</p>\",\"PeriodicalId\":73551,\"journal\":{\"name\":\"JMIR AI\",\"volume\":\"4 \",\"pages\":\"e70001\"},\"PeriodicalIF\":2.0000,\"publicationDate\":\"2025-08-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12375799/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"JMIR AI\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.2196/70001\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"JMIR AI","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2196/70001","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

背景:在线抑郁症筛查工具可能会增加循证护理的吸收,从而导致症状减轻。然而,DISCOVER试验的结果表明,与在线抑郁筛查后没有反馈相比,自动结果反馈在筛查后6个月对抑郁症状减轻没有影响。然而,在症状表征、医疗保健需求和治疗偏好方面的人际差异可能导致个体对反馈模式的不同反应。目的:本研究的目的是检验治疗效果(HTE)的异质性,即在线抑郁症筛查后对两种反馈模式(定制或非定制)与无反馈(对照)的差异反应。方法:我们使用因果森林,这是一种应用递归划分来估计条件平均处理效果(CATEs)的机器学习方法。在DISCOVER试验的次要数据分析中,符合条件的参与者至少在中度抑郁严重程度上筛选呈阳性,但在前一年未被诊断或治疗过抑郁症。主要结果是抑郁严重程度变化的异质性,在6个月的随访期间,用患者健康问卷-9测量。分析包括探索平均治疗效果(ATE), HTE,与目标操作员特征曲线(AUTOC)下的面积操作,以及根据预测的CATE分配反馈时ATE的差异。我们提取了抑郁症严重程度变化的主要预测因子,给出了反馈,并探索了高cate协变量特征。在分析之前,数据被分成训练集和测试集(1:1),以尽量减少过度拟合的风险,并评估测试数据中的预测。结果:对946名DISCOVER试验参与者的数据进行了分析,无数据缺失。我们没有检测到HTE;无定制反馈与非定制反馈,AUTOC -0.48 (95% CI -1.62至0.67,P= 0.41);无反馈与定制反馈相比,AUTOC为0.06 (95% CI -1.21至1.33,P= 0.93);无反馈vs有反馈,AUTOC为-0.20 (95% CI为-1.30 ~ 0.89,P= 0.72)。当根据预测的CATE分配反馈(定制或非定制)时,测试集中的ATE没有改变的证据。通过检查协变量概况,我们观察到,与没有反馈相比,有反馈的控制信念具有潜在的有害作用。结论:我们应用因果森林来描述大范围预测因子之间更高水平的相互作用,以检测HTE。在缺乏HTE证据的情况下,基于训练模型的治疗优先级并没有改善ATEs。我们没有发现在6个月后抑郁严重程度变化的在线抑郁筛查后提供定制或非定制反馈的危害或益处的证据。考虑到观察到的组间一致的变化,未来的研究可能会测试单独筛查是否会促进行为激活和下游抑郁严重程度的降低。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

Heterogeneity in Effects of Automated Results Feedback After Online Depression Screening: Secondary Machine-Learning Based Analysis of the DISCOVER Trial.

Heterogeneity in Effects of Automated Results Feedback After Online Depression Screening: Secondary Machine-Learning Based Analysis of the DISCOVER Trial.

Heterogeneity in Effects of Automated Results Feedback After Online Depression Screening: Secondary Machine-Learning Based Analysis of the DISCOVER Trial.

Heterogeneity in Effects of Automated Results Feedback After Online Depression Screening: Secondary Machine-Learning Based Analysis of the DISCOVER Trial.

Background: Online depression screening tools may increase uptake of evidence-based care and consequently lead to symptom reduction. However, results of the DISCOVER trial suggested no effect of automated results feedback compared with no feedback after online depression screening on depressive symptom reduction six months after screening. Interpersonal variation in symptom representation, health care needs, and treatment preferences may nonetheless have led to differential response to feedback mode on an individual level.

Objective: The aim of this study was to examine heterogeneity of treatment effects (HTE), that is, differential responses to two feedback modes (tailored or nontailored) versus no feedback (control) following online depression screening.

Methods: We used causal forests, a machine learning method that applies recursive partitioning to estimate conditional average treatment effects (CATEs). In this secondary data analysis of the DISCOVER trial, eligible participants screened positive for at least moderate depression severity but had not been diagnosed or treated for depression in the preceding year. The primary outcome was heterogeneity in depression severity change, over a and six-month follow up period, measured with the Patient Health Questionnaire-9. Analysis comprised exploration of average treatment effects (ATE), HTE, operationalized with the area under the targeting operator characteristic curve (AUTOC), and differences in ATE when allocating feedback based on predicted CATE. We extracted top predictors of depression severity change, given feedback and explored high-CATE covariate profiles. Prior to analysis, data was split into training and test sets (1:1) to minimize the risk of overfitting and evaluate predictions in held-out test data.

Results: Data from 946 participants of the DISCOVER trial without missing data were analyzed. We did not detect HTE; no versus nontailored feedback, AUTOC -0.48 (95% CI -1.62 to 0.67, P=.41); no versus tailored feedback, AUTOC 0.06 (95% CI -1.21 to 1.33, P=.93); and no versus any feedback, AUTOC -0.20 (95% CI -1.30 to 0.89, P=.72). There was no evidence of alteration to the ATE in the test set when allocating feedback (tailored or nontailored) based on the predicted CATE. By examining covariate profiles, we observed a potentially detrimental role of control beliefs, given feedback compared with no feedback.

Conclusions: We applied causal forests to describe higher-level interactions among a broad range of predictors to detect HTE. In absence of evidence for HTE, treatment prioritization based on trained models did not improve ATEs. We did not find evidence of harm or benefit from providing tailored or nontailored feedback after online depression screening regarding depression severity change after six months. Future studies may test whether screening alone prompts behavioral activation and downstream depression severity reduction, considering the observed uniform changes across groups.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信