{"title":"利用组织病理学图像对用于生存分析的视觉编码器进行基准测试","authors":"Asad Nizami, Arita Halder","doi":"10.1101/2024.08.23.24312362","DOIUrl":null,"url":null,"abstract":"Cancer is a complex disease characterized by the uncontrolled growth of abnormal cells in the body but can be prevented and even cured when detected early. Advanced medical imaging has introduced Whole Slide Images (WSIs). When combined with deep learning techniques, it can be used to extract meaningful features. These features are useful for various tasks such as classification and segmentation. There have been numerous studies involving the use of WSIs for survival analysis. Hence, it is crucial to determine their effectiveness for specific use cases. In this paper, we compared three publicly available vision encoders- UNI, Phikon and ResNet18 which are trained on millions of histopathological images, to generate feature embedding for survival analysis. WSIs cannot be fed directly to a network due to their size. We have divided them into 256 × 256 pixels patches and used a vision encoder to get feature embeddings. These embeddings were passed into an aggregator function to get representation at the WSI level which was then passed to a Long Short Term Memory (LSTM) based risk prediction head for survival analysis. Using breast cancer data from The Cancer Genome Atlas Program (TCGA) and k-fold cross-validation, we demonstrated that transformer-based models are more effective in survival analysis and achieved better C-index on average than ResNet-based architecture. The code for this study will be made available.","PeriodicalId":501437,"journal":{"name":"medRxiv - Oncology","volume":"63 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"BENCHMARKING VISION ENCODERS FOR SURVIVAL ANALYSIS USING HISTOPATHOLOGICAL IMAGES\",\"authors\":\"Asad Nizami, Arita Halder\",\"doi\":\"10.1101/2024.08.23.24312362\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Cancer is a complex disease characterized by the uncontrolled growth of abnormal cells in the body but can be prevented and even cured when detected early. Advanced medical imaging has introduced Whole Slide Images (WSIs). When combined with deep learning techniques, it can be used to extract meaningful features. These features are useful for various tasks such as classification and segmentation. There have been numerous studies involving the use of WSIs for survival analysis. Hence, it is crucial to determine their effectiveness for specific use cases. In this paper, we compared three publicly available vision encoders- UNI, Phikon and ResNet18 which are trained on millions of histopathological images, to generate feature embedding for survival analysis. WSIs cannot be fed directly to a network due to their size. We have divided them into 256 × 256 pixels patches and used a vision encoder to get feature embeddings. These embeddings were passed into an aggregator function to get representation at the WSI level which was then passed to a Long Short Term Memory (LSTM) based risk prediction head for survival analysis. Using breast cancer data from The Cancer Genome Atlas Program (TCGA) and k-fold cross-validation, we demonstrated that transformer-based models are more effective in survival analysis and achieved better C-index on average than ResNet-based architecture. The code for this study will be made available.\",\"PeriodicalId\":501437,\"journal\":{\"name\":\"medRxiv - Oncology\",\"volume\":\"63 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"medRxiv - Oncology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1101/2024.08.23.24312362\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"medRxiv - Oncology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1101/2024.08.23.24312362","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
癌症是一种复杂的疾病,其特征是异常细胞在体内不受控制地生长,但如果及早发现,是可以预防甚至治愈的。先进的医学成像技术引入了全切片图像(WSI)。当与深度学习技术相结合时,可用于提取有意义的特征。这些特征对分类和分割等各种任务非常有用。已有大量研究涉及将 WSIs 用于生存分析。因此,确定它们在特定用例中的有效性至关重要。在本文中,我们比较了三种公开可用的视觉编码器--UNI、Phikon 和 ResNet18,它们都是在数百万张组织病理学图像上经过训练的,可生成用于生存分析的特征嵌入。由于 WSIs 的大小,无法将其直接输入网络。我们将其划分为 256 × 256 像素的斑块,并使用视觉编码器获取特征嵌入。这些嵌入信息被输入聚合器函数,以获得 WSI 级别的表示,然后将其输入基于长短期记忆(LSTM)的风险预测头,用于生存分析。利用癌症基因组图谱计划(TCGA)中的乳腺癌数据和 k 倍交叉验证,我们证明了基于变换器的模型在生存分析中更为有效,平均 C 指数也优于基于 ResNet 的架构。我们将提供这项研究的代码。
BENCHMARKING VISION ENCODERS FOR SURVIVAL ANALYSIS USING HISTOPATHOLOGICAL IMAGES
Cancer is a complex disease characterized by the uncontrolled growth of abnormal cells in the body but can be prevented and even cured when detected early. Advanced medical imaging has introduced Whole Slide Images (WSIs). When combined with deep learning techniques, it can be used to extract meaningful features. These features are useful for various tasks such as classification and segmentation. There have been numerous studies involving the use of WSIs for survival analysis. Hence, it is crucial to determine their effectiveness for specific use cases. In this paper, we compared three publicly available vision encoders- UNI, Phikon and ResNet18 which are trained on millions of histopathological images, to generate feature embedding for survival analysis. WSIs cannot be fed directly to a network due to their size. We have divided them into 256 × 256 pixels patches and used a vision encoder to get feature embeddings. These embeddings were passed into an aggregator function to get representation at the WSI level which was then passed to a Long Short Term Memory (LSTM) based risk prediction head for survival analysis. Using breast cancer data from The Cancer Genome Atlas Program (TCGA) and k-fold cross-validation, we demonstrated that transformer-based models are more effective in survival analysis and achieved better C-index on average than ResNet-based architecture. The code for this study will be made available.