Saeka Rahman;Md Motiur Rahman;Miad Faezipour;Mo Rastgaar;Elika Ridelman;Justin D. Klein;Beth A. Angst;Christina M. Shanti
{"title":"Enhancing Burn Severity Assessment With Deep Learning: A Comparative Analysis and Computational Efficiency Evaluation","authors":"Saeka Rahman;Md Motiur Rahman;Miad Faezipour;Mo Rastgaar;Elika Ridelman;Justin D. Klein;Beth A. Angst;Christina M. Shanti","doi":"10.1109/ACCESS.2024.3476110","DOIUrl":null,"url":null,"abstract":"Burn injuries present a substantial global public health challenge. The conventional approach, relying on visual inspection to compute total body surface area (TBSA) for assessing burn severity, encounters the inherent limitations of proper estimation. These limitations prompted the development of computer-based applications, particularly machine learning, and deep learning models, to enhance the performance. This paper presents a comprehensive analytical study of eight deep learning techniques designed for assessing burn severity in terms of four characteristics: inflammation, scar, uniformity, and pigmentation, in small datasets of 2-dimensional (2D) images captured using digital (smartphone) camera. The models are Convolutional Neural Network (CNN), attention-based CNN, decision-level fusion (DF) based on CNN models, DF with attention-based CNN models, autoencoder-NN (Neural Network), and hybrid VGG16-Machine Learning (ML) with Support Vector Machine (SVM), Random Forest (RF), and eXtreme Gradient Boosting (XGBoost). Each model is validated with eight datasets collected and annotated by our team at the Children’s Hospital of Michigan in two phases to classify the severity of burns in terms of inflammation, scar, uniformity, and pigmentation. The average test accuracy across the eight datasets using CNN, attention-based CNN, DF with CNN models, DF with attention-based CNN models, autoencoder-NN, VGG16-RF, VGG16-SVM, and VGG16-XGBoost are \n<inline-formula> <tex-math>$0.87\\pm 0.04$ </tex-math></inline-formula>\n, \n<inline-formula> <tex-math>$0.93\\pm 0.04$ </tex-math></inline-formula>\n, \n<inline-formula> <tex-math>$0.90\\pm 0.01$ </tex-math></inline-formula>\n, \n<inline-formula> <tex-math>$0.95\\pm 0.02$ </tex-math></inline-formula>\n, \n<inline-formula> <tex-math>$0.87\\pm 0.03$ </tex-math></inline-formula>\n, \n<inline-formula> <tex-math>$0.63\\pm 0.03$ </tex-math></inline-formula>\n, \n<inline-formula> <tex-math>$0.79\\pm 0.02$ </tex-math></inline-formula>\n, \n<inline-formula> <tex-math>$0.79\\pm 0.01$ </tex-math></inline-formula>\n, correspondingly. The research also computes and compares the computational complexity of each model in terms of FLoating point Operations Per Second (FLOPS) and Multiply-ACcumulate operations (MACs). Compared with the base CNN model, the decision-level fusion with attention mechanism model outperforms with a gain of 9.19% in test accuracy and an increase of 3321.53% in FLOPS. Considering the priority and constraint of the task, the attention-based CNN model can also be preferable as it achieves an accuracy gain of 6.90% and significantly less computational increase expense (8.62%) compared with the base CNN. The code for the best performing decision-level fusion with attention mechanism model is provided on GitHub link at \n<uri>https://github.com/Saeka2022/Burn-Assessment</uri>\n.","PeriodicalId":13079,"journal":{"name":"IEEE Access","volume":"12 ","pages":"147249-147268"},"PeriodicalIF":3.4000,"publicationDate":"2024-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10707626","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Access","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10707626/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Burn injuries present a substantial global public health challenge. The conventional approach, relying on visual inspection to compute total body surface area (TBSA) for assessing burn severity, encounters the inherent limitations of proper estimation. These limitations prompted the development of computer-based applications, particularly machine learning, and deep learning models, to enhance the performance. This paper presents a comprehensive analytical study of eight deep learning techniques designed for assessing burn severity in terms of four characteristics: inflammation, scar, uniformity, and pigmentation, in small datasets of 2-dimensional (2D) images captured using digital (smartphone) camera. The models are Convolutional Neural Network (CNN), attention-based CNN, decision-level fusion (DF) based on CNN models, DF with attention-based CNN models, autoencoder-NN (Neural Network), and hybrid VGG16-Machine Learning (ML) with Support Vector Machine (SVM), Random Forest (RF), and eXtreme Gradient Boosting (XGBoost). Each model is validated with eight datasets collected and annotated by our team at the Children’s Hospital of Michigan in two phases to classify the severity of burns in terms of inflammation, scar, uniformity, and pigmentation. The average test accuracy across the eight datasets using CNN, attention-based CNN, DF with CNN models, DF with attention-based CNN models, autoencoder-NN, VGG16-RF, VGG16-SVM, and VGG16-XGBoost are
$0.87\pm 0.04$
,
$0.93\pm 0.04$
,
$0.90\pm 0.01$
,
$0.95\pm 0.02$
,
$0.87\pm 0.03$
,
$0.63\pm 0.03$
,
$0.79\pm 0.02$
,
$0.79\pm 0.01$
, correspondingly. The research also computes and compares the computational complexity of each model in terms of FLoating point Operations Per Second (FLOPS) and Multiply-ACcumulate operations (MACs). Compared with the base CNN model, the decision-level fusion with attention mechanism model outperforms with a gain of 9.19% in test accuracy and an increase of 3321.53% in FLOPS. Considering the priority and constraint of the task, the attention-based CNN model can also be preferable as it achieves an accuracy gain of 6.90% and significantly less computational increase expense (8.62%) compared with the base CNN. The code for the best performing decision-level fusion with attention mechanism model is provided on GitHub link at
https://github.com/Saeka2022/Burn-Assessment
.
IEEE AccessCOMPUTER SCIENCE, INFORMATION SYSTEMSENGIN-ENGINEERING, ELECTRICAL & ELECTRONIC
CiteScore
9.80
自引率
7.70%
发文量
6673
审稿时长
6 weeks
期刊介绍:
IEEE Access® is a multidisciplinary, open access (OA), applications-oriented, all-electronic archival journal that continuously presents the results of original research or development across all of IEEE''s fields of interest.
IEEE Access will publish articles that are of high interest to readers, original, technically correct, and clearly presented. Supported by author publication charges (APC), its hallmarks are a rapid peer review and publication process with open access to all readers. Unlike IEEE''s traditional Transactions or Journals, reviews are "binary", in that reviewers will either Accept or Reject an article in the form it is submitted in order to achieve rapid turnaround. Especially encouraged are submissions on:
Multidisciplinary topics, or applications-oriented articles and negative results that do not fit within the scope of IEEE''s traditional journals.
Practical articles discussing new experiments or measurement techniques, interesting solutions to engineering.
Development of new or improved fabrication or manufacturing techniques.
Reviews or survey articles of new or evolving fields oriented to assist others in understanding the new area.