{"title":"Pre-trained vs. Random Weights for Calculating Fréchet Inception Distance in Medical Imaging","authors":"Jamie A. O’Reilly, Fawad Asadi","doi":"10.1109/BMEiCON53485.2021.9745214","DOIUrl":null,"url":null,"abstract":"Fréchet Inception Distance (FID) is an evaluation metric for assessing the quality of images synthesized by generative models. Conventionally this involves using an Inception v3 convolutional neural network that has been pretrained to classify everyday color images with the ImageNet dataset. The final classification section of this network is omitted, leaving an efficient feature extractor that outputs an encoded representation of each input image in the form of a 2048 element vector. Difference or similarity between samples of images can then be compared by measuring the distance between the distributions of their corresponding feature representations. Researchers have raised concerns about the utility of FID for evaluating unorthodox images (e.g. medical images) that are unlike those used for model training; suggesting that randomly initialized convolutional neural networks may be more appropriate. The aim of this study was to compare pre-trained and random approaches for evaluating medical images. Robustness to synthetic image distortions (Gaussian noise, blurring, swirl, and impulse noise) and different image types (Xray, CT, fundus, and everyday images) was addressed. Feature representations were converted into two-dimensional space and visualized using t-distributed stochastic neighbor embedding (tSNE) and principal component analysis (PCA). Normalized FID between image classes was substantially larger and more consistent for the pre-trained model. Overall, this suggests that the pre-trained model is preferable to the randomly initialized model for evaluating medical images.","PeriodicalId":380002,"journal":{"name":"2021 13th Biomedical Engineering International Conference (BMEiCON)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 13th Biomedical Engineering International Conference (BMEiCON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/BMEiCON53485.2021.9745214","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Fréchet Inception Distance (FID) is an evaluation metric for assessing the quality of images synthesized by generative models. Conventionally this involves using an Inception v3 convolutional neural network that has been pretrained to classify everyday color images with the ImageNet dataset. The final classification section of this network is omitted, leaving an efficient feature extractor that outputs an encoded representation of each input image in the form of a 2048 element vector. Difference or similarity between samples of images can then be compared by measuring the distance between the distributions of their corresponding feature representations. Researchers have raised concerns about the utility of FID for evaluating unorthodox images (e.g. medical images) that are unlike those used for model training; suggesting that randomly initialized convolutional neural networks may be more appropriate. The aim of this study was to compare pre-trained and random approaches for evaluating medical images. Robustness to synthetic image distortions (Gaussian noise, blurring, swirl, and impulse noise) and different image types (Xray, CT, fundus, and everyday images) was addressed. Feature representations were converted into two-dimensional space and visualized using t-distributed stochastic neighbor embedding (tSNE) and principal component analysis (PCA). Normalized FID between image classes was substantially larger and more consistent for the pre-trained model. Overall, this suggests that the pre-trained model is preferable to the randomly initialized model for evaluating medical images.