Automated segmentation of child-clinician speech in naturalistic clinical contexts

IF 2.9 2区 医学 Q1 EDUCATION, SPECIAL
Giulio Bertamini , Cesare Furlanello , Mohamed Chetouani , David Cohen , Paola Venuti
{"title":"Automated segmentation of child-clinician speech in naturalistic clinical contexts","authors":"Giulio Bertamini ,&nbsp;Cesare Furlanello ,&nbsp;Mohamed Chetouani ,&nbsp;David Cohen ,&nbsp;Paola Venuti","doi":"10.1016/j.ridd.2024.104906","DOIUrl":null,"url":null,"abstract":"<div><h3>Background</h3><div>Computational approaches hold significant promise for enhancing diagnosis and therapy in child and adolescent clinical practice. Clinical procedures heavily depend n vocal exchanges and interpersonal dynamics conveyed through speech. Research highlights the importance of investigating acoustic features and dyadic interactions during child development. However, observational methods are labor-intensive, time-consuming, and suffer from limited objectivity and quantification, hindering translation to everyday care.</div></div><div><h3>Aims</h3><div>We propose a novel AI-based system for fully automatic acoustic segmentation of clinical sessions with autistic preschool children.</div></div><div><h3>Methods and procedures</h3><div>We focused on naturalistic and unconstrained clinical contexts, which are characterized by background noise and data scarcity. Our approach addresses key challenges in the field while remaining non-invasive. We carefully evaluated model performance and flexibility in diverse, challenging conditions by means of domain alignment.</div></div><div><h3>Outcomes and results</h3><div>Results demonstrated promising outcomes in voice activity detection and speaker diarization. Notably, minimal annotation efforts —just 30 seconds of target data— significantly improved model performance across all tested conditions. Our models exhibit satisfying predictive performance and flexibility for deployment in everyday settings.</div></div><div><h3>Conclusions and implications</h3><div>Automating data annotation in real-world clinical scenarios can enable the widespread exploitation of advanced computational methods for downstream modeling, fostering precision approaches that bridge research and clinical practice.</div></div>","PeriodicalId":51351,"journal":{"name":"Research in Developmental Disabilities","volume":"157 ","pages":"Article 104906"},"PeriodicalIF":2.9000,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Research in Developmental Disabilities","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0891422224002385","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"EDUCATION, SPECIAL","Score":null,"Total":0}
引用次数: 0

Abstract

Background

Computational approaches hold significant promise for enhancing diagnosis and therapy in child and adolescent clinical practice. Clinical procedures heavily depend n vocal exchanges and interpersonal dynamics conveyed through speech. Research highlights the importance of investigating acoustic features and dyadic interactions during child development. However, observational methods are labor-intensive, time-consuming, and suffer from limited objectivity and quantification, hindering translation to everyday care.

Aims

We propose a novel AI-based system for fully automatic acoustic segmentation of clinical sessions with autistic preschool children.

Methods and procedures

We focused on naturalistic and unconstrained clinical contexts, which are characterized by background noise and data scarcity. Our approach addresses key challenges in the field while remaining non-invasive. We carefully evaluated model performance and flexibility in diverse, challenging conditions by means of domain alignment.

Outcomes and results

Results demonstrated promising outcomes in voice activity detection and speaker diarization. Notably, minimal annotation efforts —just 30 seconds of target data— significantly improved model performance across all tested conditions. Our models exhibit satisfying predictive performance and flexibility for deployment in everyday settings.

Conclusions and implications

Automating data annotation in real-world clinical scenarios can enable the widespread exploitation of advanced computational methods for downstream modeling, fostering precision approaches that bridge research and clinical practice.
自然临床语境中儿童临床医生语音的自动分割。
背景:计算方法在提高儿童和青少年临床实践的诊断和治疗方面具有重要的前景。临床程序在很大程度上依赖于通过言语传达的声音交换和人际动态。研究强调了在儿童发育过程中调查声学特征和二元相互作用的重要性。然而,观察方法劳动密集,耗时,客观性和量化有限,阻碍了转化为日常护理。目的:我们提出了一种新的基于人工智能的孤独症学龄前儿童临床会话全自动声学分割系统。方法和程序:我们关注自然和不受约束的临床环境,其特点是背景噪声和数据稀缺。我们的方法解决了该领域的关键挑战,同时保持了非侵入性。我们通过领域对齐的方式仔细评估了模型在各种具有挑战性的条件下的性能和灵活性。结果和结果:结果显示在语音活动检测和说话人拨号方面有希望的结果。值得注意的是,最小的注释工作(目标数据只需30 秒)显著提高了模型在所有测试条件下的性能。我们的模型表现出令人满意的预测性能和在日常环境中部署的灵活性。结论和启示:在现实世界的临床场景中自动化数据注释可以广泛利用先进的计算方法进行下游建模,促进精确的方法,桥梁研究和临床实践。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
5.50
自引率
6.50%
发文量
178
期刊介绍: Research In Developmental Disabilities is aimed at publishing original research of an interdisciplinary nature that has a direct bearing on the remediation of problems associated with developmental disabilities. Manuscripts will be solicited throughout the world. Articles will be primarily empirical studies, although an occasional position paper or review will be accepted. The aim of the journal will be to publish articles on all aspects of research with the developmentally disabled, with any methodologically sound approach being acceptable.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信