{"title":"袋内采样对端到端多实例学习的影响","authors":"N. Koriakina, Natasa Sladoje, Joakim Lindblad","doi":"10.1109/ISPA52656.2021.9552170","DOIUrl":null,"url":null,"abstract":"End-to-end multiple instance learning (MIL) is an important concept with a wide range of applications. It is gaining increased popularity in the (bio)medical imaging community since it may provide a possibility to, while relying only on weak labels assigned to large regions, obtain more fine-grained information. However, processing very large bags in end-to-end MIL is problematic due to computer memory constraints. We propose within-bag sampling as one way of applying end-to-end MIL methods on very large data. We explore how different levels of sampling affect the performance of a well-known high-performing end-to-end attention-based MIL method, to understand the conditions when sampling can be utilized. We compose two new datasets tailored for the purpose of the study, and propose a strategy for sampling during MIL inference to arrive at reliable bag labels as well as instance level attention weights. We perform experiments without and with different levels of sampling, on the two publicly available datasets, and for a range of learning settings. We observe that in most situations the proposed bag-level sampling can be applied to end-to-end MIL without performance loss, supporting its confident usage to enable end-to-end MIL also in scenarios with very large bags. We share the code as open source at https://github.com/MIDA-group/SampledABMIL","PeriodicalId":131088,"journal":{"name":"2021 12th International Symposium on Image and Signal Processing and Analysis (ISPA)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"The Effect of Within-Bag Sampling on End-to-End Multiple Instance Learning\",\"authors\":\"N. Koriakina, Natasa Sladoje, Joakim Lindblad\",\"doi\":\"10.1109/ISPA52656.2021.9552170\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"End-to-end multiple instance learning (MIL) is an important concept with a wide range of applications. It is gaining increased popularity in the (bio)medical imaging community since it may provide a possibility to, while relying only on weak labels assigned to large regions, obtain more fine-grained information. However, processing very large bags in end-to-end MIL is problematic due to computer memory constraints. We propose within-bag sampling as one way of applying end-to-end MIL methods on very large data. We explore how different levels of sampling affect the performance of a well-known high-performing end-to-end attention-based MIL method, to understand the conditions when sampling can be utilized. We compose two new datasets tailored for the purpose of the study, and propose a strategy for sampling during MIL inference to arrive at reliable bag labels as well as instance level attention weights. We perform experiments without and with different levels of sampling, on the two publicly available datasets, and for a range of learning settings. We observe that in most situations the proposed bag-level sampling can be applied to end-to-end MIL without performance loss, supporting its confident usage to enable end-to-end MIL also in scenarios with very large bags. We share the code as open source at https://github.com/MIDA-group/SampledABMIL\",\"PeriodicalId\":131088,\"journal\":{\"name\":\"2021 12th International Symposium on Image and Signal Processing and Analysis (ISPA)\",\"volume\":\"18 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-09-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 12th International Symposium on Image and Signal Processing and Analysis (ISPA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISPA52656.2021.9552170\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 12th International Symposium on Image and Signal Processing and Analysis (ISPA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISPA52656.2021.9552170","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The Effect of Within-Bag Sampling on End-to-End Multiple Instance Learning
End-to-end multiple instance learning (MIL) is an important concept with a wide range of applications. It is gaining increased popularity in the (bio)medical imaging community since it may provide a possibility to, while relying only on weak labels assigned to large regions, obtain more fine-grained information. However, processing very large bags in end-to-end MIL is problematic due to computer memory constraints. We propose within-bag sampling as one way of applying end-to-end MIL methods on very large data. We explore how different levels of sampling affect the performance of a well-known high-performing end-to-end attention-based MIL method, to understand the conditions when sampling can be utilized. We compose two new datasets tailored for the purpose of the study, and propose a strategy for sampling during MIL inference to arrive at reliable bag labels as well as instance level attention weights. We perform experiments without and with different levels of sampling, on the two publicly available datasets, and for a range of learning settings. We observe that in most situations the proposed bag-level sampling can be applied to end-to-end MIL without performance loss, supporting its confident usage to enable end-to-end MIL also in scenarios with very large bags. We share the code as open source at https://github.com/MIDA-group/SampledABMIL