A medical image classification method based on self-regularized adversarial learning

IF 3.2 2区 医学 Q1 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING
Medical physics Pub Date : 2024-07-30 DOI:10.1002/mp.17320
Zong Fan, Xiaohui Zhang, Su Ruan, Wade Thorstad, Hiram Gay, Pengfei Song, Xiaowei Wang, Hua Li
{"title":"A medical image classification method based on self-regularized adversarial learning","authors":"Zong Fan,&nbsp;Xiaohui Zhang,&nbsp;Su Ruan,&nbsp;Wade Thorstad,&nbsp;Hiram Gay,&nbsp;Pengfei Song,&nbsp;Xiaowei Wang,&nbsp;Hua Li","doi":"10.1002/mp.17320","DOIUrl":null,"url":null,"abstract":"<div>\n \n \n <section>\n \n <h3> Background</h3>\n \n <p>Deep learning (DL) techniques have been extensively applied in medical image classification. The unique characteristics of medical imaging data present challenges, including small labeled datasets, severely imbalanced class distribution, and significant variations in imaging quality. Recently, generative adversarial network (GAN)-based classification methods have gained attention for their ability to enhance classification accuracy by incorporating realistic GAN-generated images as data augmentation. However, the performance of these GAN-based methods often relies on high-quality generated images, while large amounts of training data are required to train GAN models to achieve optimal performance.</p>\n </section>\n \n <section>\n \n <h3> Purpose</h3>\n \n <p>In this study, we propose an adversarial learning-based classification framework to achieve better classification performance. Innovatively, GAN models are employed as supplementary regularization terms to support classification, aiming to address the challenges described above.</p>\n </section>\n \n <section>\n \n <h3> Methods</h3>\n \n <p>The proposed classification framework, GAN-DL, consists of a feature extraction network (F-Net), a classifier, and two adversarial networks, specifically a reconstruction network (R-Net) and a discriminator network (D-Net). The F-Net extracts features from input images, and the classifier uses these features for classification tasks. R-Net and D-Net have been designed following the GAN architecture. R-Net employs the extracted feature to reconstruct the original images, while D-Net is tasked with the discrimination between the reconstructed image and the original images. An iterative adversarial learning strategy is designed to guide model training by incorporating multiple network-specific loss functions. These loss functions, serving as supplementary regularization, are automatically derived during the reconstruction process and require no additional data annotation.</p>\n </section>\n \n <section>\n \n <h3> Results</h3>\n \n <p>To verify the model's effectiveness, we performed experiments on two datasets, including a COVID-19 dataset with 13 958 chest x-ray images and an oropharyngeal squamous cell carcinoma (OPSCC) dataset with 3255 positron emission tomography images. Thirteen classic DL-based classification methods were implemented on the same datasets for comparison. Performance metrics included precision, sensitivity, specificity, and <span></span><math>\n <semantics>\n <msub>\n <mi>F</mi>\n <mn>1</mn>\n </msub>\n <annotation>$F_1$</annotation>\n </semantics></math>-score. In addition, we conducted ablation studies to assess the effects of various factors on model performance, including the network depth of F-Net, training image size, training dataset size, and loss function design. Our method achieved superior performance than all comparative methods. On the COVID-19 dataset, our method achieved <span></span><math>\n <semantics>\n <mrow>\n <mn>95.4</mn>\n <mo>%</mo>\n <mo>±</mo>\n <mn>0.6</mn>\n <mo>%</mo>\n </mrow>\n <annotation>$95.4\\%\\pm 0.6\\%$</annotation>\n </semantics></math>, <span></span><math>\n <semantics>\n <mrow>\n <mn>95.3</mn>\n <mo>%</mo>\n <mo>±</mo>\n <mn>0.9</mn>\n <mo>%</mo>\n </mrow>\n <annotation>$95.3\\%\\pm 0.9\\%$</annotation>\n </semantics></math>, <span></span><math>\n <semantics>\n <mrow>\n <mn>97.7</mn>\n <mo>%</mo>\n <mo>±</mo>\n <mn>0.4</mn>\n <mo>%</mo>\n </mrow>\n <annotation>$97.7\\%\\pm 0.4\\%$</annotation>\n </semantics></math>, and <span></span><math>\n <semantics>\n <mrow>\n <mn>95.3</mn>\n <mo>%</mo>\n <mo>±</mo>\n <mn>0.9</mn>\n <mo>%</mo>\n </mrow>\n <annotation>$95.3\\%\\pm 0.9\\%$</annotation>\n </semantics></math> in terms of precision, sensitivity, specificity, and <span></span><math>\n <semantics>\n <msub>\n <mi>F</mi>\n <mn>1</mn>\n </msub>\n <annotation>$F_1$</annotation>\n </semantics></math>-score, respectively. It achieved <span></span><math>\n <semantics>\n <mrow>\n <mn>96.2</mn>\n <mo>%</mo>\n <mo>±</mo>\n <mn>0.7</mn>\n <mo>%</mo>\n </mrow>\n <annotation>$96.2\\%\\pm 0.7\\%$</annotation>\n </semantics></math> across all these metrics on the OPSCC dataset. The study to investigate the effects of two adversarial networks highlights the crucial role of D-Net in improving model performance. Ablation studies further provide an in-depth understanding of our methodology.</p>\n </section>\n \n <section>\n \n <h3> Conclusion</h3>\n \n <p>Our adversarial-based classification framework leverages GAN-based adversarial networks and an iterative adversarial learning strategy to harness supplementary regularization during training. This design significantly enhances classification accuracy and mitigates overfitting issues in medical image datasets. Moreover, its modular design not only demonstrates flexibility but also indicates its potential applicability to various clinical contexts and medical imaging applications.</p>\n </section>\n </div>","PeriodicalId":18384,"journal":{"name":"Medical physics","volume":"51 11","pages":"8232-8246"},"PeriodicalIF":3.2000,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical physics","FirstCategoryId":"3","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/mp.17320","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0

Abstract

Background

Deep learning (DL) techniques have been extensively applied in medical image classification. The unique characteristics of medical imaging data present challenges, including small labeled datasets, severely imbalanced class distribution, and significant variations in imaging quality. Recently, generative adversarial network (GAN)-based classification methods have gained attention for their ability to enhance classification accuracy by incorporating realistic GAN-generated images as data augmentation. However, the performance of these GAN-based methods often relies on high-quality generated images, while large amounts of training data are required to train GAN models to achieve optimal performance.

Purpose

In this study, we propose an adversarial learning-based classification framework to achieve better classification performance. Innovatively, GAN models are employed as supplementary regularization terms to support classification, aiming to address the challenges described above.

Methods

The proposed classification framework, GAN-DL, consists of a feature extraction network (F-Net), a classifier, and two adversarial networks, specifically a reconstruction network (R-Net) and a discriminator network (D-Net). The F-Net extracts features from input images, and the classifier uses these features for classification tasks. R-Net and D-Net have been designed following the GAN architecture. R-Net employs the extracted feature to reconstruct the original images, while D-Net is tasked with the discrimination between the reconstructed image and the original images. An iterative adversarial learning strategy is designed to guide model training by incorporating multiple network-specific loss functions. These loss functions, serving as supplementary regularization, are automatically derived during the reconstruction process and require no additional data annotation.

Results

To verify the model's effectiveness, we performed experiments on two datasets, including a COVID-19 dataset with 13 958 chest x-ray images and an oropharyngeal squamous cell carcinoma (OPSCC) dataset with 3255 positron emission tomography images. Thirteen classic DL-based classification methods were implemented on the same datasets for comparison. Performance metrics included precision, sensitivity, specificity, and F 1 $F_1$ -score. In addition, we conducted ablation studies to assess the effects of various factors on model performance, including the network depth of F-Net, training image size, training dataset size, and loss function design. Our method achieved superior performance than all comparative methods. On the COVID-19 dataset, our method achieved 95.4 % ± 0.6 % $95.4\%\pm 0.6\%$ , 95.3 % ± 0.9 % $95.3\%\pm 0.9\%$ , 97.7 % ± 0.4 % $97.7\%\pm 0.4\%$ , and 95.3 % ± 0.9 % $95.3\%\pm 0.9\%$ in terms of precision, sensitivity, specificity, and F 1 $F_1$ -score, respectively. It achieved 96.2 % ± 0.7 % $96.2\%\pm 0.7\%$ across all these metrics on the OPSCC dataset. The study to investigate the effects of two adversarial networks highlights the crucial role of D-Net in improving model performance. Ablation studies further provide an in-depth understanding of our methodology.

Conclusion

Our adversarial-based classification framework leverages GAN-based adversarial networks and an iterative adversarial learning strategy to harness supplementary regularization during training. This design significantly enhances classification accuracy and mitigates overfitting issues in medical image datasets. Moreover, its modular design not only demonstrates flexibility but also indicates its potential applicability to various clinical contexts and medical imaging applications.

基于自规范对抗学习的医学图像分类方法。
背景:深度学习(DL)技术已被广泛应用于医学影像分类。医学影像数据的独特性带来了挑战,包括标记数据集小、类分布严重失衡以及成像质量差异大。最近,基于生成式对抗网络(GAN)的分类方法受到了关注,因为这些方法能够通过将 GAN 生成的真实图像作为数据增强来提高分类准确性。然而,这些基于生成式对抗网络的方法的性能往往依赖于高质量的生成图像,同时需要大量的训练数据来训练生成式对抗网络模型以达到最佳性能。创新性地采用 GAN 模型作为支持分类的辅助正则化条件,旨在解决上述挑战:所提出的分类框架 GAN-DL 包括一个特征提取网络(F-Net)、一个分类器和两个对抗网络,特别是一个重构网络(R-Net)和一个判别网络(D-Net)。F-Net 从输入图像中提取特征,分类器利用这些特征完成分类任务。R-Net 和 D-Net 是按照 GAN 架构设计的。R-Net 利用提取的特征重建原始图像,而 D-Net 的任务是区分重建图像和原始图像。我们设计了一种迭代对抗学习策略,通过纳入多个特定网络的损失函数来指导模型训练。这些损失函数作为补充正则化,在重建过程中自动得出,无需额外的数据注释:为了验证模型的有效性,我们在两个数据集上进行了实验,其中包括包含 13 958 张胸部 X 光图像的 COVID-19 数据集和含 3255 张正电子发射断层扫描图像的口咽鳞状细胞癌(OPSCC)数据集。在相同的数据集上采用了 13 种基于 DL 的经典分类方法进行比较。性能指标包括精确度、灵敏度、特异性和 F 1 $F_1$ 评分。此外,我们还进行了消融研究,以评估各种因素对模型性能的影响,包括 F-Net 的网络深度、训练图像大小、训练数据集大小和损失函数设计。与所有比较方法相比,我们的方法取得了更优越的性能。在 COVID-19 数据集上,我们的方法取得了 95.4 % ± 0.6 % $95.4\pm 0.6\$ 、 95.3 % ± 0.9 % $95.3\pm 0.9\$ 、 97.7 % ± 0.4 % $97.在精确度、灵敏度、特异性和 F 1 $F_1$ -score 方面分别为 95.3 % ± 0.9 % $95.3 % ± 0.9 % $97.7 % ± 0.4 % $97.7 % ± 0.4 % $97.3 % ± 0.9 % $95.3 % ± 0.9 % $95.3%/pm0.9%$。在 OPSCC 数据集上,它在所有这些指标上都达到了 96.2 % ± 0.7 % $96.2%/pm 0.7%$。对两个对抗网络效果的研究突出了 D-Net 在提高模型性能方面的关键作用。消融研究进一步深入了解了我们的方法:我们基于对抗的分类框架利用基于 GAN 的对抗网络和迭代对抗学习策略,在训练过程中利用补充正则化。这种设计大大提高了分类准确性,缓解了医学图像数据集的过拟合问题。此外,它的模块化设计不仅体现了灵活性,还表明了它在各种临床环境和医学影像应用中的潜在适用性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Medical physics
Medical physics 医学-核医学
CiteScore
6.80
自引率
15.80%
发文量
660
审稿时长
1.7 months
期刊介绍: Medical Physics publishes original, high impact physics, imaging science, and engineering research that advances patient diagnosis and therapy through contributions in 1) Basic science developments with high potential for clinical translation 2) Clinical applications of cutting edge engineering and physics innovations 3) Broadly applicable and innovative clinical physics developments Medical Physics is a journal of global scope and reach. By publishing in Medical Physics your research will reach an international, multidisciplinary audience including practicing medical physicists as well as physics- and engineering based translational scientists. We work closely with authors of promising articles to improve their quality.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信