Hongjian Song, Jie Tang, Hongzhao Xiao, Juncheng Hu
{"title":"Rethinking Overfitting of Multiple Instance Learning for Whole Slide Image Classification","authors":"Hongjian Song, Jie Tang, Hongzhao Xiao, Juncheng Hu","doi":"10.1109/ICME55011.2023.00100","DOIUrl":null,"url":null,"abstract":"Multiple instance learning(MIL) is widely used for whole slide image(WSI) classification. However, these methods suffer from severe overfitting. In this paper, we introduce two main causes of such overfitting problems by rethinking the MIL task and formulation of attention-based MIL models: (i) The model is sensitive to the proportion of positive regions, and (ii)incorrectly learns the positional relationship of patches (i.e., the order of instances). To this end, we propose recurrent random padding(RRP) module and patch shuffle(PS) module to tackle these two issues, respectively. Furthermore, we present random alignment(RA) algorithm to solve these two overfitting problems simultaneously. On CAMELYON16 and TCGA-NSCLC, the proposed plug-and-play modules improve the performance of six baselines by large margins. The significant and consistent refinement demonstrates the correctness of our theories and the effectiveness of our modules.","PeriodicalId":321830,"journal":{"name":"2023 IEEE International Conference on Multimedia and Expo (ICME)","volume":"66 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE International Conference on Multimedia and Expo (ICME)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICME55011.2023.00100","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Multiple instance learning(MIL) is widely used for whole slide image(WSI) classification. However, these methods suffer from severe overfitting. In this paper, we introduce two main causes of such overfitting problems by rethinking the MIL task and formulation of attention-based MIL models: (i) The model is sensitive to the proportion of positive regions, and (ii)incorrectly learns the positional relationship of patches (i.e., the order of instances). To this end, we propose recurrent random padding(RRP) module and patch shuffle(PS) module to tackle these two issues, respectively. Furthermore, we present random alignment(RA) algorithm to solve these two overfitting problems simultaneously. On CAMELYON16 and TCGA-NSCLC, the proposed plug-and-play modules improve the performance of six baselines by large margins. The significant and consistent refinement demonstrates the correctness of our theories and the effectiveness of our modules.