Zac E Imel, Brian Pace, Brad Pendergraft, Jordan Pruett, Michael Tanana, Christina S Soma, Kate A Comtois, David C Atkins
{"title":"基于机器学习的危机咨询电话自杀风险评估。","authors":"Zac E Imel, Brian Pace, Brad Pendergraft, Jordan Pruett, Michael Tanana, Christina S Soma, Kate A Comtois, David C Atkins","doi":"10.1176/appi.ps.20230648","DOIUrl":null,"url":null,"abstract":"<p><strong>Objective: </strong>Counselor assessment of suicide risk is one key component of crisis counseling, and standards require risk assessment in every crisis counseling conversation. Efforts to increase risk assessment frequency are limited by quality improvement tools that rely on human evaluation of conversations, which is labor intensive, slow, and impossible to scale. Advances in machine learning (ML) have made possible the development of tools that can automatically and immediately detect the presence of risk assessment in crisis counseling conversations.</p><p><strong>Methods: </strong>To train models, a coding team labeled every statement in 476 crisis counseling calls (193,257 statements) for a core element of risk assessment. The authors then fine-tuned a transformer-based ML model with the labeled data, utilizing separate training, validation, and test data sets.</p><p><strong>Results: </strong>Generally, the evaluated ML model was highly consistent with human raters. For detecting any risk assessment, ML model agreement with human ratings was 98% of human interrater agreement. Across specific labels, average F1 (the harmonic mean of precision and recall) was 0.86 at the call level and 0.66 at the statement level and often varied as a result of a low base rate for some risk labels.</p><p><strong>Conclusions: </strong>ML models can reliably detect the presence of suicide risk assessment in crisis counseling conversations, presenting an opportunity to scale quality improvement efforts.</p>","PeriodicalId":20878,"journal":{"name":"Psychiatric services","volume":null,"pages":null},"PeriodicalIF":3.3000,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11530329/pdf/","citationCount":"0","resultStr":"{\"title\":\"Machine Learning-Based Evaluation of Suicide Risk Assessment in Crisis Counseling Calls.\",\"authors\":\"Zac E Imel, Brian Pace, Brad Pendergraft, Jordan Pruett, Michael Tanana, Christina S Soma, Kate A Comtois, David C Atkins\",\"doi\":\"10.1176/appi.ps.20230648\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Objective: </strong>Counselor assessment of suicide risk is one key component of crisis counseling, and standards require risk assessment in every crisis counseling conversation. Efforts to increase risk assessment frequency are limited by quality improvement tools that rely on human evaluation of conversations, which is labor intensive, slow, and impossible to scale. Advances in machine learning (ML) have made possible the development of tools that can automatically and immediately detect the presence of risk assessment in crisis counseling conversations.</p><p><strong>Methods: </strong>To train models, a coding team labeled every statement in 476 crisis counseling calls (193,257 statements) for a core element of risk assessment. The authors then fine-tuned a transformer-based ML model with the labeled data, utilizing separate training, validation, and test data sets.</p><p><strong>Results: </strong>Generally, the evaluated ML model was highly consistent with human raters. For detecting any risk assessment, ML model agreement with human ratings was 98% of human interrater agreement. Across specific labels, average F1 (the harmonic mean of precision and recall) was 0.86 at the call level and 0.66 at the statement level and often varied as a result of a low base rate for some risk labels.</p><p><strong>Conclusions: </strong>ML models can reliably detect the presence of suicide risk assessment in crisis counseling conversations, presenting an opportunity to scale quality improvement efforts.</p>\",\"PeriodicalId\":20878,\"journal\":{\"name\":\"Psychiatric services\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":3.3000,\"publicationDate\":\"2024-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11530329/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Psychiatric services\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1176/appi.ps.20230648\",\"RegionNum\":3,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2024/7/19 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"Q1\",\"JCRName\":\"HEALTH POLICY & SERVICES\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Psychiatric services","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1176/appi.ps.20230648","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/7/19 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"HEALTH POLICY & SERVICES","Score":null,"Total":0}
引用次数: 0
摘要
目的:心理咨询师对自杀风险的评估是危机咨询的一个关键组成部分,标准要求在每次危机咨询谈话中都要进行风险评估。提高风险评估频率的努力受到了质量改进工具的限制,这些工具依赖于对谈话的人工评估,而人工评估耗费人力、速度慢,而且无法扩大规模。机器学习(ML)的进步使得开发工具成为可能,这些工具可以自动并立即检测危机咨询对话中是否存在风险评估:为了训练模型,一个编码团队对 476 个危机咨询电话(193,257 个语句)中的每个语句都标注了风险评估的核心要素。然后,作者利用单独的训练、验证和测试数据集,对基于转换器的 ML 模型进行了微调:总体而言,经过评估的 ML 模型与人类评定者高度一致。在检测任何风险评估时,ML 模型与人类评分的一致性是人类评分者之间一致性的 98%。在所有具体标签中,平均 F1(精确度和召回率的调和平均值)在通话级别为 0.86,在语句级别为 0.66,并且由于某些风险标签的基准率较低而经常变化:结论:ML 模型可以可靠地检测危机咨询对话中是否存在自杀风险评估,为提高咨询质量提供了机会。
Machine Learning-Based Evaluation of Suicide Risk Assessment in Crisis Counseling Calls.
Objective: Counselor assessment of suicide risk is one key component of crisis counseling, and standards require risk assessment in every crisis counseling conversation. Efforts to increase risk assessment frequency are limited by quality improvement tools that rely on human evaluation of conversations, which is labor intensive, slow, and impossible to scale. Advances in machine learning (ML) have made possible the development of tools that can automatically and immediately detect the presence of risk assessment in crisis counseling conversations.
Methods: To train models, a coding team labeled every statement in 476 crisis counseling calls (193,257 statements) for a core element of risk assessment. The authors then fine-tuned a transformer-based ML model with the labeled data, utilizing separate training, validation, and test data sets.
Results: Generally, the evaluated ML model was highly consistent with human raters. For detecting any risk assessment, ML model agreement with human ratings was 98% of human interrater agreement. Across specific labels, average F1 (the harmonic mean of precision and recall) was 0.86 at the call level and 0.66 at the statement level and often varied as a result of a low base rate for some risk labels.
Conclusions: ML models can reliably detect the presence of suicide risk assessment in crisis counseling conversations, presenting an opportunity to scale quality improvement efforts.
期刊介绍:
Psychiatric Services, established in 1950, is published monthly by the American Psychiatric Association. The peer-reviewed journal features research reports on issues related to the delivery of mental health services, especially for people with serious mental illness in community-based treatment programs. Long known as an interdisciplinary journal, Psychiatric Services recognizes that provision of high-quality care involves collaboration among a variety of professionals, frequently working as a team. Authors of research reports published in the journal include psychiatrists, psychologists, pharmacists, nurses, social workers, drug and alcohol treatment counselors, economists, policy analysts, and professionals in related systems such as criminal justice and welfare systems. In the mental health field, the current focus on patient-centered, recovery-oriented care and on dissemination of evidence-based practices is transforming service delivery systems at all levels. Research published in Psychiatric Services contributes to this transformation.