{"title":"GAN-based Spatial Transformation Adversarial Method for Disease Classification on CXR Photographs by Smartphones","authors":"Chak Fong Chong, Xu Yang, Wei Ke, Yapeng Wang","doi":"10.1109/DICTA52665.2021.9647192","DOIUrl":null,"url":null,"abstract":"Deep learning has been successfully applied on Chest X-ray (CXR) images for disease classification. To support remote medical services (e.g., online diagnosis services), such systems can be deployed on smartphones by patients or doctors to take CXR photographs using the cameras on smartphones. However, photograph introduces visual artifacts such as blur, noises, light reflection, perspective transformation, moiré pattern, etc. plus unwanted background. Therefore, the classification accuracy of well-trained CNN models performed on the CXR photographs experiences drop significantly. Such challenge has not been solved properly in the literature. In this paper, we have compared various traditional image preprocessing methods on CXR photographs, including spatial transformation, background hiding, and various filtering methods. The combination of these methods can almost eliminate the negative impact of visual artifacts on the evaluation of 3 different single CNN models (Xception, DenseNet-121, Inception-v3), only 0.0018 AUC drop observed. However, such methods need user manually process the CXR photographs, which is inconvenient. Therefore, we have proposed a novel Generative Adversarial Network-based spatial transformation adversarial method (GAN-STAM) which can automatically transform the CXR region to the center and enlarge the CXR region in each CXR photograph, the classification accuracy has been significantly improved on CXR photographs from 0.8009 to 0.8653.","PeriodicalId":424950,"journal":{"name":"2021 Digital Image Computing: Techniques and Applications (DICTA)","volume":"117 2","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 Digital Image Computing: Techniques and Applications (DICTA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DICTA52665.2021.9647192","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
Deep learning has been successfully applied on Chest X-ray (CXR) images for disease classification. To support remote medical services (e.g., online diagnosis services), such systems can be deployed on smartphones by patients or doctors to take CXR photographs using the cameras on smartphones. However, photograph introduces visual artifacts such as blur, noises, light reflection, perspective transformation, moiré pattern, etc. plus unwanted background. Therefore, the classification accuracy of well-trained CNN models performed on the CXR photographs experiences drop significantly. Such challenge has not been solved properly in the literature. In this paper, we have compared various traditional image preprocessing methods on CXR photographs, including spatial transformation, background hiding, and various filtering methods. The combination of these methods can almost eliminate the negative impact of visual artifacts on the evaluation of 3 different single CNN models (Xception, DenseNet-121, Inception-v3), only 0.0018 AUC drop observed. However, such methods need user manually process the CXR photographs, which is inconvenient. Therefore, we have proposed a novel Generative Adversarial Network-based spatial transformation adversarial method (GAN-STAM) which can automatically transform the CXR region to the center and enlarge the CXR region in each CXR photograph, the classification accuracy has been significantly improved on CXR photographs from 0.8009 to 0.8653.