Sunghwa You , Chanbeom Kwak , Chul Young Yoon , Young Joon Seo
{"title":"用深度学习数字化听力图:结构化数据提取和假名化听觉大数据","authors":"Sunghwa You , Chanbeom Kwak , Chul Young Yoon , Young Joon Seo","doi":"10.1016/j.heares.2025.109337","DOIUrl":null,"url":null,"abstract":"<div><h3>Purpose</h3><div>hearing loss relies on pure-tone audiometry (PTA); however, audiograms are often stored as unstructured images, limiting their integration into electronic medical records (EMRs) and common data models (CDMs). This study developed a deep learning-based system to digitize audiograms, enabling the structured and numerical conversion of data for large-scale hearing big data collection.</div></div><div><h3>Methods</h3><div>A convolutional neural network (CNN) was trained to extract numerical frequency and threshold values from audiograms. The system consists of four modules: preprocessing, pattern classification, image analysis, and post-processing. Optical character recognition (OCR) was employed to extract patient data, which were then pseudonymized to prevent leakage of personal and sensitive information. The model was trained using 8847 audiometric symbols and tested using 2443 symbols.</div></div><div><h3>Results</h3><div>The model achieved accuracy of 95.01 % and 98.18 % for the right and left ears, respectively. It processed audiograms 17.72 times faster than manual digitization, reducing processing time from 63.27 s to 3.57 s per audiogram. The structured data format allows seamless integration into big data and CDMs, ensuring compliance with pseudonymization and anonymization protocols.</div></div><div><h3>Discussion</h3><div>The model improves data accessibility and scalability for both clinical and research applications. Unlike previous studies that primarily focused on classification or prediction, this framework ensures a structured numerical data output while adhering to data pseudonymization regulations.</div></div><div><h3>Conclusion</h3><div>This deep learning-based system enhanced the efficiency and accuracy of audiogram digitization, facilitating the construction of hearing big data, integration with CDMs, AI-driven diagnostics, and large-scale hearing data analysis.</div></div>","PeriodicalId":12881,"journal":{"name":"Hearing Research","volume":"464 ","pages":"Article 109337"},"PeriodicalIF":2.5000,"publicationDate":"2025-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Digitizing audiograms with deep learning: structured data extraction and pseudonymization for hearing big data\",\"authors\":\"Sunghwa You , Chanbeom Kwak , Chul Young Yoon , Young Joon Seo\",\"doi\":\"10.1016/j.heares.2025.109337\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><h3>Purpose</h3><div>hearing loss relies on pure-tone audiometry (PTA); however, audiograms are often stored as unstructured images, limiting their integration into electronic medical records (EMRs) and common data models (CDMs). This study developed a deep learning-based system to digitize audiograms, enabling the structured and numerical conversion of data for large-scale hearing big data collection.</div></div><div><h3>Methods</h3><div>A convolutional neural network (CNN) was trained to extract numerical frequency and threshold values from audiograms. The system consists of four modules: preprocessing, pattern classification, image analysis, and post-processing. Optical character recognition (OCR) was employed to extract patient data, which were then pseudonymized to prevent leakage of personal and sensitive information. The model was trained using 8847 audiometric symbols and tested using 2443 symbols.</div></div><div><h3>Results</h3><div>The model achieved accuracy of 95.01 % and 98.18 % for the right and left ears, respectively. It processed audiograms 17.72 times faster than manual digitization, reducing processing time from 63.27 s to 3.57 s per audiogram. The structured data format allows seamless integration into big data and CDMs, ensuring compliance with pseudonymization and anonymization protocols.</div></div><div><h3>Discussion</h3><div>The model improves data accessibility and scalability for both clinical and research applications. Unlike previous studies that primarily focused on classification or prediction, this framework ensures a structured numerical data output while adhering to data pseudonymization regulations.</div></div><div><h3>Conclusion</h3><div>This deep learning-based system enhanced the efficiency and accuracy of audiogram digitization, facilitating the construction of hearing big data, integration with CDMs, AI-driven diagnostics, and large-scale hearing data analysis.</div></div>\",\"PeriodicalId\":12881,\"journal\":{\"name\":\"Hearing Research\",\"volume\":\"464 \",\"pages\":\"Article 109337\"},\"PeriodicalIF\":2.5000,\"publicationDate\":\"2025-06-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Hearing Research\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0378595525001558\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Hearing Research","FirstCategoryId":"3","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0378595525001558","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY","Score":null,"Total":0}
Digitizing audiograms with deep learning: structured data extraction and pseudonymization for hearing big data
Purpose
hearing loss relies on pure-tone audiometry (PTA); however, audiograms are often stored as unstructured images, limiting their integration into electronic medical records (EMRs) and common data models (CDMs). This study developed a deep learning-based system to digitize audiograms, enabling the structured and numerical conversion of data for large-scale hearing big data collection.
Methods
A convolutional neural network (CNN) was trained to extract numerical frequency and threshold values from audiograms. The system consists of four modules: preprocessing, pattern classification, image analysis, and post-processing. Optical character recognition (OCR) was employed to extract patient data, which were then pseudonymized to prevent leakage of personal and sensitive information. The model was trained using 8847 audiometric symbols and tested using 2443 symbols.
Results
The model achieved accuracy of 95.01 % and 98.18 % for the right and left ears, respectively. It processed audiograms 17.72 times faster than manual digitization, reducing processing time from 63.27 s to 3.57 s per audiogram. The structured data format allows seamless integration into big data and CDMs, ensuring compliance with pseudonymization and anonymization protocols.
Discussion
The model improves data accessibility and scalability for both clinical and research applications. Unlike previous studies that primarily focused on classification or prediction, this framework ensures a structured numerical data output while adhering to data pseudonymization regulations.
Conclusion
This deep learning-based system enhanced the efficiency and accuracy of audiogram digitization, facilitating the construction of hearing big data, integration with CDMs, AI-driven diagnostics, and large-scale hearing data analysis.
期刊介绍:
The aim of the journal is to provide a forum for papers concerned with basic peripheral and central auditory mechanisms. Emphasis is on experimental and clinical studies, but theoretical and methodological papers will also be considered. The journal publishes original research papers, review and mini- review articles, rapid communications, method/protocol and perspective articles.
Papers submitted should deal with auditory anatomy, physiology, psychophysics, imaging, modeling and behavioural studies in animals and humans, as well as hearing aids and cochlear implants. Papers dealing with the vestibular system are also considered for publication. Papers on comparative aspects of hearing and on effects of drugs and environmental contaminants on hearing function will also be considered. Clinical papers will be accepted when they contribute to the understanding of normal and pathological hearing functions.