{"title":"使用基于深度学习的对象检测开发一种听力图解释诊断支持系统。","authors":"Titipat Achakulvisut, Suchanon Phanthong, Thanawut Timpitak, Kanpat Vesessook, Sirinan Junthong, Withita Utainrat, Kanokrat Bunnag","doi":"10.26599/JOTO.2025.9540005","DOIUrl":null,"url":null,"abstract":"<p><strong>Objective: </strong>To develop and evaluate an automated system for digitizing audiograms, classifying hearing loss levels, and comparing their performance with traditional methods and otolaryngologists' interpretations.</p><p><strong>Designed and methods: </strong>We conducted a retrospective diagnostic study using 1,959 audiogram images from patients aged 7 years and older at the Faculty of Medicine, Vajira Hospital, Navamindradhiraj University. We employed an object detection approach to digitize audiograms and developed multiple machine learning models to classify six hearing loss levels. The dataset was split into 70% training (1,407 images) and 30% testing (352 images) sets. We compared our model's performance with classifications based on manually extracted audiogram values and otolaryngologists' interpretations.</p><p><strong>Result: </strong>Our object detection-based model achieved an F1-score of 94.72% in classifying hearing loss levels, comparable to the 96.43% F1-score obtained using manually extracted values. The Light Gradient Boosting Machine (LGBM) model is used as the classifier for the manually extracted data, which achieved top performance with 94.72% accuracy, 94.72% f1-score, 94.72 recall, and 94.72 precision. In object detection based model, The Random Forest Classifier (RFC) model showed the highest 96.43% accuracy in predicting hearing loss level, with a F1-score of 96.43%, recall of 96.43%, and precision of 96.45%.</p><p><strong>Conclusion: </strong>Our proposed automated approach for audiogram digitization and hearing loss classification performs comparably to traditional methods and otolaryngologists' interpretations. This system can potentially assist otolaryngologists in providing more timely and effective treatment by quickly and accurately classifying hearing loss.</p>","PeriodicalId":94336,"journal":{"name":"Journal of otology","volume":"20 1","pages":"26-32"},"PeriodicalIF":0.0000,"publicationDate":"2025-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12510341/pdf/","citationCount":"0","resultStr":"{\"title\":\"Developing a diagnostic support system for audiogram interpretation using deep learning-based object detection.\",\"authors\":\"Titipat Achakulvisut, Suchanon Phanthong, Thanawut Timpitak, Kanpat Vesessook, Sirinan Junthong, Withita Utainrat, Kanokrat Bunnag\",\"doi\":\"10.26599/JOTO.2025.9540005\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Objective: </strong>To develop and evaluate an automated system for digitizing audiograms, classifying hearing loss levels, and comparing their performance with traditional methods and otolaryngologists' interpretations.</p><p><strong>Designed and methods: </strong>We conducted a retrospective diagnostic study using 1,959 audiogram images from patients aged 7 years and older at the Faculty of Medicine, Vajira Hospital, Navamindradhiraj University. We employed an object detection approach to digitize audiograms and developed multiple machine learning models to classify six hearing loss levels. The dataset was split into 70% training (1,407 images) and 30% testing (352 images) sets. We compared our model's performance with classifications based on manually extracted audiogram values and otolaryngologists' interpretations.</p><p><strong>Result: </strong>Our object detection-based model achieved an F1-score of 94.72% in classifying hearing loss levels, comparable to the 96.43% F1-score obtained using manually extracted values. The Light Gradient Boosting Machine (LGBM) model is used as the classifier for the manually extracted data, which achieved top performance with 94.72% accuracy, 94.72% f1-score, 94.72 recall, and 94.72 precision. In object detection based model, The Random Forest Classifier (RFC) model showed the highest 96.43% accuracy in predicting hearing loss level, with a F1-score of 96.43%, recall of 96.43%, and precision of 96.45%.</p><p><strong>Conclusion: </strong>Our proposed automated approach for audiogram digitization and hearing loss classification performs comparably to traditional methods and otolaryngologists' interpretations. This system can potentially assist otolaryngologists in providing more timely and effective treatment by quickly and accurately classifying hearing loss.</p>\",\"PeriodicalId\":94336,\"journal\":{\"name\":\"Journal of otology\",\"volume\":\"20 1\",\"pages\":\"26-32\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-03-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12510341/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of otology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.26599/JOTO.2025.9540005\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/1/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of otology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.26599/JOTO.2025.9540005","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/1 0:00:00","PubModel":"eCollection","JCR":"","JCRName":"","Score":null,"Total":0}
Developing a diagnostic support system for audiogram interpretation using deep learning-based object detection.
Objective: To develop and evaluate an automated system for digitizing audiograms, classifying hearing loss levels, and comparing their performance with traditional methods and otolaryngologists' interpretations.
Designed and methods: We conducted a retrospective diagnostic study using 1,959 audiogram images from patients aged 7 years and older at the Faculty of Medicine, Vajira Hospital, Navamindradhiraj University. We employed an object detection approach to digitize audiograms and developed multiple machine learning models to classify six hearing loss levels. The dataset was split into 70% training (1,407 images) and 30% testing (352 images) sets. We compared our model's performance with classifications based on manually extracted audiogram values and otolaryngologists' interpretations.
Result: Our object detection-based model achieved an F1-score of 94.72% in classifying hearing loss levels, comparable to the 96.43% F1-score obtained using manually extracted values. The Light Gradient Boosting Machine (LGBM) model is used as the classifier for the manually extracted data, which achieved top performance with 94.72% accuracy, 94.72% f1-score, 94.72 recall, and 94.72 precision. In object detection based model, The Random Forest Classifier (RFC) model showed the highest 96.43% accuracy in predicting hearing loss level, with a F1-score of 96.43%, recall of 96.43%, and precision of 96.45%.
Conclusion: Our proposed automated approach for audiogram digitization and hearing loss classification performs comparably to traditional methods and otolaryngologists' interpretations. This system can potentially assist otolaryngologists in providing more timely and effective treatment by quickly and accurately classifying hearing loss.