A foundation model for generalizable cancer diagnosis and survival prediction from histopathological images

Zhangsheng Yu, Zhaochang Yang, Ting Wei, Ying Liang, Xin Yuan, Ruitian Gao, Yujia Xia, Jie Zhou, Yue Zhang
{"title":"A foundation model for generalizable cancer diagnosis and survival prediction from histopathological images","authors":"Zhangsheng Yu, Zhaochang Yang, Ting Wei, Ying Liang, Xin Yuan, Ruitian Gao, Yujia Xia, Jie Zhou, Yue Zhang","doi":"10.1101/2024.05.16.594499","DOIUrl":null,"url":null,"abstract":"Computational pathology, utilizing whole slide image (WSI) for pathological diagnosis, has advanced the development of intelligent healthcare. However, the scarcity of annotated data and histological differences hinder the general application of existing methods. Extensive\nhistopathological data and the robustness of self-supervised models in small-scale data demonstrate promising prospects for developing foundation pathology models. In this work, we propose the BEPH (BEiT-based model Pre-training on Histopathological image), a general method that leverages self-supervised learning to learn meaningful representations from 11 million unlabeled\nhistopathological images. These representations are then efficiently adapted to various tasks, including patch-level cancer recognition, WSI-level cancer classification, and survival prediction for multiple cancer subtypes. Experimental results demonstrate that our model consistently outperformsseveral comparative models, even with limited training data reduced to 50%. Especially\nwhen the downstream structure is the same, the model can improve ResNet and DINO by up to a maximum increase of 8.8% and 7.2% (WSI level classification), and 6.44% and 3.28% on average (survival prediction), respectively. Therefore, BEPH offers a universal solution to enhance model performance, reduce the burden of expert annotations, and enable widespread clinical applications of artificial intelligence. The code and models can be obtained at https://github.com/Zhcyoung/BEPH. And currently, online fine-tuning of WSI classification tasks\nis available for use on http://yulab-sjtu.natapp1.cc/BEPH.","PeriodicalId":501471,"journal":{"name":"bioRxiv - Pathology","volume":"66 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"bioRxiv - Pathology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1101/2024.05.16.594499","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Computational pathology, utilizing whole slide image (WSI) for pathological diagnosis, has advanced the development of intelligent healthcare. However, the scarcity of annotated data and histological differences hinder the general application of existing methods. Extensive histopathological data and the robustness of self-supervised models in small-scale data demonstrate promising prospects for developing foundation pathology models. In this work, we propose the BEPH (BEiT-based model Pre-training on Histopathological image), a general method that leverages self-supervised learning to learn meaningful representations from 11 million unlabeled histopathological images. These representations are then efficiently adapted to various tasks, including patch-level cancer recognition, WSI-level cancer classification, and survival prediction for multiple cancer subtypes. Experimental results demonstrate that our model consistently outperformsseveral comparative models, even with limited training data reduced to 50%. Especially when the downstream structure is the same, the model can improve ResNet and DINO by up to a maximum increase of 8.8% and 7.2% (WSI level classification), and 6.44% and 3.28% on average (survival prediction), respectively. Therefore, BEPH offers a universal solution to enhance model performance, reduce the burden of expert annotations, and enable widespread clinical applications of artificial intelligence. The code and models can be obtained at https://github.com/Zhcyoung/BEPH. And currently, online fine-tuning of WSI classification tasks is available for use on http://yulab-sjtu.natapp1.cc/BEPH.
根据组织病理图像进行癌症诊断和生存预测的基础模型
利用全切片图像(WSI)进行病理诊断的计算病理学推动了智能医疗的发展。然而,注释数据的匮乏和组织学差异阻碍了现有方法的普遍应用。广泛的病理数据和自监督模型在小规模数据中的鲁棒性为开发基础病理模型展示了广阔的前景。在这项工作中,我们提出了 BEPH(基于 BEiT 的组织病理学图像预训练模型),这是一种利用自我监督学习从 1100 万张无标签组织病理学图像中学习有意义表征的通用方法。然后,这些表征被有效地应用于各种任务,包括斑块级癌症识别、WSI 级癌症分类和多种癌症亚型的生存预测。实验结果表明,即使将有限的训练数据减少到 50%,我们的模型也始终优于其他比较模型。尤其是在下游结构相同的情况下,该模型对 ResNet 和 DINO 的改进最大可达 8.8% 和 7.2%(WSI 级分类),平均 6.44% 和 3.28%(生存预测)。因此,BEPH 为提高模型性能、减轻专家注释负担以及实现人工智能的广泛临床应用提供了通用解决方案。代码和模型可从 https://github.com/Zhcyoung/BEPH 获取。目前,WSI 分类任务的在线微调可在 http://yulab-sjtu.natapp1.cc/BEPH 上使用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信