Sonia Baee, Jeremy W Eberle, Anna N Baglione, Tyler Spears, Elijah Lewis, Hongning Wang, Daniel H Funk, Bethany Teachman, Laura E Barnes
{"title":"基于网络的解释偏差修正的早期流失预测,以减少焦虑思维:机器学习研究。","authors":"Sonia Baee, Jeremy W Eberle, Anna N Baglione, Tyler Spears, Elijah Lewis, Hongning Wang, Daniel H Funk, Bethany Teachman, Laura E Barnes","doi":"10.2196/51567","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Digital mental health is a promising paradigm for individualized, patient-driven health care. For example, cognitive bias modification programs that target interpretation biases (cognitive bias modification for interpretation [CBM-I]) can provide practice thinking about ambiguous situations in less threatening ways on the web without requiring a therapist. However, digital mental health interventions, including CBM-I, are often plagued with lack of sustained engagement and high attrition rates. New attrition detection and mitigation strategies are needed to improve these interventions.</p><p><strong>Objective: </strong>This paper aims to identify participants at a high risk of dropout during the early stages of 3 web-based trials of multisession CBM-I and to investigate which self-reported and passively detected feature sets computed from the participants interacting with the intervention and assessments were most informative in making this prediction.</p><p><strong>Methods: </strong>The participants analyzed in this paper were community adults with traits such as anxiety or negative thinking about the future (Study 1: n=252, Study 2: n=326, Study 3: n=699) who had been assigned to CBM-I conditions in 3 efficacy-effectiveness trials on our team's public research website. To identify participants at a high risk of dropout, we created 4 unique feature sets: self-reported baseline user characteristics (eg, demographics), self-reported user context and reactions to the program (eg, state affect), self-reported user clinical functioning (eg, mental health symptoms), and passively detected user behavior on the website (eg, time spent on a web page of CBM-I training exercises, time of day during which the exercises were completed, latency of completing the assessments, and type of device used). Then, we investigated the feature sets as potential predictors of which participants were at high risk of not starting the second training session of a given program using well-known machine learning algorithms.</p><p><strong>Results: </strong>The extreme gradient boosting algorithm performed the best and identified participants at high risk with macro-F<sub>1</sub>-scores of .832 (Study 1 with 146 features), .770 (Study 2 with 87 features), and .917 (Study 3 with 127 features). Features involving passive detection of user behavior contributed the most to the prediction relative to other features. The mean Gini importance scores for the passive features were as follows: .033 (95% CI .019-.047) in Study 1; .029 (95% CI .023-.035) in Study 2; and .045 (95% CI .039-.051) in Study 3. However, using all features extracted from a given study led to the best predictive performance.</p><p><strong>Conclusions: </strong>These results suggest that using passive indicators of user behavior, alongside self-reported measures, can improve the accuracy of prediction of participants at a high risk of dropout early during multisession CBM-I programs. Furthermore, our analyses highlight the challenge of generalizability in digital health intervention studies and the need for more personalized attrition prevention strategies.</p>","PeriodicalId":48616,"journal":{"name":"Jmir Mental Health","volume":"11 ","pages":"e51567"},"PeriodicalIF":4.8000,"publicationDate":"2024-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11699492/pdf/","citationCount":"0","resultStr":"{\"title\":\"Early Attrition Prediction for Web-Based Interpretation Bias Modification to Reduce Anxious Thinking: A Machine Learning Study.\",\"authors\":\"Sonia Baee, Jeremy W Eberle, Anna N Baglione, Tyler Spears, Elijah Lewis, Hongning Wang, Daniel H Funk, Bethany Teachman, Laura E Barnes\",\"doi\":\"10.2196/51567\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Background: </strong>Digital mental health is a promising paradigm for individualized, patient-driven health care. For example, cognitive bias modification programs that target interpretation biases (cognitive bias modification for interpretation [CBM-I]) can provide practice thinking about ambiguous situations in less threatening ways on the web without requiring a therapist. However, digital mental health interventions, including CBM-I, are often plagued with lack of sustained engagement and high attrition rates. New attrition detection and mitigation strategies are needed to improve these interventions.</p><p><strong>Objective: </strong>This paper aims to identify participants at a high risk of dropout during the early stages of 3 web-based trials of multisession CBM-I and to investigate which self-reported and passively detected feature sets computed from the participants interacting with the intervention and assessments were most informative in making this prediction.</p><p><strong>Methods: </strong>The participants analyzed in this paper were community adults with traits such as anxiety or negative thinking about the future (Study 1: n=252, Study 2: n=326, Study 3: n=699) who had been assigned to CBM-I conditions in 3 efficacy-effectiveness trials on our team's public research website. To identify participants at a high risk of dropout, we created 4 unique feature sets: self-reported baseline user characteristics (eg, demographics), self-reported user context and reactions to the program (eg, state affect), self-reported user clinical functioning (eg, mental health symptoms), and passively detected user behavior on the website (eg, time spent on a web page of CBM-I training exercises, time of day during which the exercises were completed, latency of completing the assessments, and type of device used). Then, we investigated the feature sets as potential predictors of which participants were at high risk of not starting the second training session of a given program using well-known machine learning algorithms.</p><p><strong>Results: </strong>The extreme gradient boosting algorithm performed the best and identified participants at high risk with macro-F<sub>1</sub>-scores of .832 (Study 1 with 146 features), .770 (Study 2 with 87 features), and .917 (Study 3 with 127 features). Features involving passive detection of user behavior contributed the most to the prediction relative to other features. The mean Gini importance scores for the passive features were as follows: .033 (95% CI .019-.047) in Study 1; .029 (95% CI .023-.035) in Study 2; and .045 (95% CI .039-.051) in Study 3. However, using all features extracted from a given study led to the best predictive performance.</p><p><strong>Conclusions: </strong>These results suggest that using passive indicators of user behavior, alongside self-reported measures, can improve the accuracy of prediction of participants at a high risk of dropout early during multisession CBM-I programs. Furthermore, our analyses highlight the challenge of generalizability in digital health intervention studies and the need for more personalized attrition prevention strategies.</p>\",\"PeriodicalId\":48616,\"journal\":{\"name\":\"Jmir Mental Health\",\"volume\":\"11 \",\"pages\":\"e51567\"},\"PeriodicalIF\":4.8000,\"publicationDate\":\"2024-12-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11699492/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Jmir Mental Health\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.2196/51567\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"PSYCHIATRY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Jmir Mental Health","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.2196/51567","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHIATRY","Score":null,"Total":0}
引用次数: 0
摘要
背景:数字心理健康是个性化、患者驱动的医疗保健的一个很有前途的范例。例如,针对解释偏见的认知偏见修正程序(解释的认知偏见修正[CBM-I])可以在不需要治疗师的情况下,以不那么具有威胁性的方式在网络上提供对模糊情况的思考练习。然而,数字心理健康干预措施,包括CBM-I,往往受到缺乏持续参与和高流失率的困扰。需要新的磨损检测和缓解战略来改进这些干预措施。目的:本文旨在确定在3个基于网络的多阶段CBM-I试验的早期阶段处于高风险的参与者,并调查从参与者与干预和评估相互作用中计算出的自我报告和被动检测特征集在做出这一预测时最具信息性。方法:本文分析的参与者是具有焦虑或对未来消极思考等特征的社区成年人(研究1:n=252,研究2:n=326,研究3:n=699),他们在我们团队的公共研究网站上进行了3次疗效试验,被分配到CBM-I条件。为了识别退学风险高的参与者,我们创建了4个独特的特征集:自我报告的基线用户特征(例如,人口统计数据)、自我报告的用户背景和对程序的反应(例如,状态影响)、自我报告的用户临床功能(例如,心理健康症状),以及被动检测到的用户在网站上的行为(例如,在CBM-I训练练习的网页上花费的时间、完成练习的时间、完成评估的延迟时间和使用的设备类型)。然后,我们研究了特征集作为潜在的预测因素,哪些参与者在使用知名机器学习算法的给定程序中不开始第二次训练的风险很高。结果:极端梯度增强算法表现最好,识别出高风险参与者的宏观f1得分为0.832(研究1有146个特征)、0.770(研究2有87个特征)和0.917(研究3有127个特征)。相对于其他特征,涉及被动检测用户行为的特征对预测的贡献最大。被动特征的平均基尼重要性评分如下:研究1中为0.033 (95% CI为0.019 - 0.047);研究2中为0.029 (95% CI 0.023 - 0.035);研究3和0.045 (95% CI 0.039 - 0.051)。然而,使用从给定研究中提取的所有特征会导致最佳的预测性能。结论:这些结果表明,使用被动的用户行为指标,以及自我报告的措施,可以提高预测多会话CBM-I计划中早期退学高风险参与者的准确性。此外,我们的分析强调了数字健康干预研究中普遍性的挑战,以及对更个性化的磨损预防策略的需求。
Early Attrition Prediction for Web-Based Interpretation Bias Modification to Reduce Anxious Thinking: A Machine Learning Study.
Background: Digital mental health is a promising paradigm for individualized, patient-driven health care. For example, cognitive bias modification programs that target interpretation biases (cognitive bias modification for interpretation [CBM-I]) can provide practice thinking about ambiguous situations in less threatening ways on the web without requiring a therapist. However, digital mental health interventions, including CBM-I, are often plagued with lack of sustained engagement and high attrition rates. New attrition detection and mitigation strategies are needed to improve these interventions.
Objective: This paper aims to identify participants at a high risk of dropout during the early stages of 3 web-based trials of multisession CBM-I and to investigate which self-reported and passively detected feature sets computed from the participants interacting with the intervention and assessments were most informative in making this prediction.
Methods: The participants analyzed in this paper were community adults with traits such as anxiety or negative thinking about the future (Study 1: n=252, Study 2: n=326, Study 3: n=699) who had been assigned to CBM-I conditions in 3 efficacy-effectiveness trials on our team's public research website. To identify participants at a high risk of dropout, we created 4 unique feature sets: self-reported baseline user characteristics (eg, demographics), self-reported user context and reactions to the program (eg, state affect), self-reported user clinical functioning (eg, mental health symptoms), and passively detected user behavior on the website (eg, time spent on a web page of CBM-I training exercises, time of day during which the exercises were completed, latency of completing the assessments, and type of device used). Then, we investigated the feature sets as potential predictors of which participants were at high risk of not starting the second training session of a given program using well-known machine learning algorithms.
Results: The extreme gradient boosting algorithm performed the best and identified participants at high risk with macro-F1-scores of .832 (Study 1 with 146 features), .770 (Study 2 with 87 features), and .917 (Study 3 with 127 features). Features involving passive detection of user behavior contributed the most to the prediction relative to other features. The mean Gini importance scores for the passive features were as follows: .033 (95% CI .019-.047) in Study 1; .029 (95% CI .023-.035) in Study 2; and .045 (95% CI .039-.051) in Study 3. However, using all features extracted from a given study led to the best predictive performance.
Conclusions: These results suggest that using passive indicators of user behavior, alongside self-reported measures, can improve the accuracy of prediction of participants at a high risk of dropout early during multisession CBM-I programs. Furthermore, our analyses highlight the challenge of generalizability in digital health intervention studies and the need for more personalized attrition prevention strategies.
期刊介绍:
JMIR Mental Health (JMH, ISSN 2368-7959) is a PubMed-indexed, peer-reviewed sister journal of JMIR, the leading eHealth journal (Impact Factor 2016: 5.175).
JMIR Mental Health focusses on digital health and Internet interventions, technologies and electronic innovations (software and hardware) for mental health, addictions, online counselling and behaviour change. This includes formative evaluation and system descriptions, theoretical papers, review papers, viewpoint/vision papers, and rigorous evaluations.