ResHNet: Spectrograms Based Efficient Heart Sounds Classification Using Stacked Residual Networks

A. Balamurugan, S. Teo, Jianxi Yang, Zhongbo Peng, Xulei Yang, Zeng Zeng
{"title":"ResHNet: Spectrograms Based Efficient Heart Sounds Classification Using Stacked Residual Networks","authors":"A. Balamurugan, S. Teo, Jianxi Yang, Zhongbo Peng, Xulei Yang, Zeng Zeng","doi":"10.1109/BHI.2019.8834578","DOIUrl":null,"url":null,"abstract":"Cardiovascular disease (CVD) is one of the major contributors of global mortality rate as it accounts for almost 31% of the worldwide deaths. As per World Health Organization (WHO), CVD continues to be the number one cause of death in the world. In some parts of the world, access to expert doctors and diagnosis are difficult. Thus an efficient and quick diagnosis of heart disease method is needed especially for low-income and middle-income countries where Magnetic Resonance Imaging (MRI) and Ultrasound becomes a constraint in terms of the resources to save human life. With the tremendous technology advancement in the medical field, deep learning has gained more attention to automate most of the initial diagnosis of diseases. This fosters continuous research in adopting deep learning methods for automatic classification of heart sounds to identify any abnormalities. In this work, we aim to investigate the efficiency of introducing residual modules in heart sounds classification using a deep neural network. This approach involves the following steps: (i) Generation of Spectrograms for every 1D audio signal using Spectrogram generator module (ii) Training of residual network based classifier for identifying normal and abnormal heart sounds based on the Spectrograms. The standard dataset given for 2016 PhysioNet/CinC Challenge has been used here for validating our residual network. This method achieved around 97% accuracy on the independent hidden test set performing best without incorporating any segmentation or additional Mel-Frequency Cepstrum Coefficients (MFCC) features of the audio signals and just the learned features from the image-based representations. Various baseline results of other deep learning based approaches have also been considered for evaluating the robustness of this framework.","PeriodicalId":281971,"journal":{"name":"2019 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/BHI.2019.8834578","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5

Abstract

Cardiovascular disease (CVD) is one of the major contributors of global mortality rate as it accounts for almost 31% of the worldwide deaths. As per World Health Organization (WHO), CVD continues to be the number one cause of death in the world. In some parts of the world, access to expert doctors and diagnosis are difficult. Thus an efficient and quick diagnosis of heart disease method is needed especially for low-income and middle-income countries where Magnetic Resonance Imaging (MRI) and Ultrasound becomes a constraint in terms of the resources to save human life. With the tremendous technology advancement in the medical field, deep learning has gained more attention to automate most of the initial diagnosis of diseases. This fosters continuous research in adopting deep learning methods for automatic classification of heart sounds to identify any abnormalities. In this work, we aim to investigate the efficiency of introducing residual modules in heart sounds classification using a deep neural network. This approach involves the following steps: (i) Generation of Spectrograms for every 1D audio signal using Spectrogram generator module (ii) Training of residual network based classifier for identifying normal and abnormal heart sounds based on the Spectrograms. The standard dataset given for 2016 PhysioNet/CinC Challenge has been used here for validating our residual network. This method achieved around 97% accuracy on the independent hidden test set performing best without incorporating any segmentation or additional Mel-Frequency Cepstrum Coefficients (MFCC) features of the audio signals and just the learned features from the image-based representations. Various baseline results of other deep learning based approaches have also been considered for evaluating the robustness of this framework.
ResHNet:利用堆叠残差网络,基于频谱图的高效心音分类
心血管疾病(CVD)是全球死亡率的主要原因之一,占全球死亡人数的近31%。根据世界卫生组织(WHO)的数据,心血管疾病仍然是世界上最大的死亡原因。在世界某些地区,很难获得专家医生和诊断。因此,需要一种有效和快速的心脏病诊断方法,特别是对于低收入和中等收入国家,磁共振成像(MRI)和超声在拯救人类生命的资源方面成为一种限制。随着医学领域的巨大技术进步,深度学习越来越受到人们的关注,因为它可以自动化大部分疾病的初始诊断。这促进了采用深度学习方法自动分类心音以识别任何异常的持续研究。在这项工作中,我们旨在研究使用深度神经网络在心音分类中引入残差模块的效率。该方法包括以下步骤:(i)使用谱图生成器模块为每个1D音频信号生成谱图(ii)基于谱图训练基于残差网络的分类器,用于识别正常和异常心音。这里使用了2016年PhysioNet/CinC挑战赛的标准数据集来验证我们的残差网络。该方法在独立隐藏测试集上达到了约97%的准确率,在不结合音频信号的任何分割或额外的Mel-Frequency倒频谱系数(MFCC)特征以及仅从基于图像的表示中学习的特征的情况下表现最佳。其他基于深度学习的方法的各种基线结果也被考虑用于评估该框架的鲁棒性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信