{"title":"The transfer learning gap: quantifying transfer learning in a medical image case","authors":"Javier Guerra-Librero, M. Bento, R. Frayne","doi":"10.1117/12.2670071","DOIUrl":null,"url":null,"abstract":"Transfer learning is a widely used technique in medical imaging and other research fields where a scarcity of available data limit the training of machine learning algorithms. Despite its widespread use and extensive supporting body of research, the specific mechanisms behind transfer learning are not completely understood. In this work, we quantify the effectiveness of transfer learning in medical image classification scenarios for different numbers of training set images. We trained ResNet50, a popular deep learning model used in medical image classification, using two scenarios: 1) applying transfer learning to a pre-trained network and 2) training the same model from scratch (i.e., starting with randomly selected weights). We analyzed the performance of the model under both scenarios as the number of training set images increased from 5,000 to 160,000 medical images. We introduced and evaluated a metric, the transfer learning gap (TLG), to quantify the differences between the two scenarios. The TLG measured the difference in the area under the loss curves (AULCs) when transfer learning was applied and when the model was trained from scratch. Our experiments show that as the training set size increases, the TLG trends to zero, suggesting that the advantage of using transfer learning decreases. The trend in the AULC suggests a training set size where the two scenarios would have equal losses. At this point, the model reaches the same performance regardless of if transfer learning or training from scratch was used. This study is important because it provides a novel metric to understand and quantify the effect of transfer learning.","PeriodicalId":147201,"journal":{"name":"Symposium on Medical Information Processing and Analysis","volume":"208 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Symposium on Medical Information Processing and Analysis","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.2670071","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Transfer learning is a widely used technique in medical imaging and other research fields where a scarcity of available data limit the training of machine learning algorithms. Despite its widespread use and extensive supporting body of research, the specific mechanisms behind transfer learning are not completely understood. In this work, we quantify the effectiveness of transfer learning in medical image classification scenarios for different numbers of training set images. We trained ResNet50, a popular deep learning model used in medical image classification, using two scenarios: 1) applying transfer learning to a pre-trained network and 2) training the same model from scratch (i.e., starting with randomly selected weights). We analyzed the performance of the model under both scenarios as the number of training set images increased from 5,000 to 160,000 medical images. We introduced and evaluated a metric, the transfer learning gap (TLG), to quantify the differences between the two scenarios. The TLG measured the difference in the area under the loss curves (AULCs) when transfer learning was applied and when the model was trained from scratch. Our experiments show that as the training set size increases, the TLG trends to zero, suggesting that the advantage of using transfer learning decreases. The trend in the AULC suggests a training set size where the two scenarios would have equal losses. At this point, the model reaches the same performance regardless of if transfer learning or training from scratch was used. This study is important because it provides a novel metric to understand and quantify the effect of transfer learning.