Swin transformers are robust to distribution and concept drift in endoscopy-based longitudinal rectal cancer assessment.

Jorge Tapias Gomez, Aneesh Rangnekar, Hannah Williams, Hannah M Thompson, Julio Garcia-Aguilar, Joshua Jesse Smith, Harini Veeraraghavan
{"title":"Swin transformers are robust to distribution and concept drift in endoscopy-based longitudinal rectal cancer assessment.","authors":"Jorge Tapias Gomez, Aneesh Rangnekar, Hannah Williams, Hannah M Thompson, Julio Garcia-Aguilar, Joshua Jesse Smith, Harini Veeraraghavan","doi":"10.1117/12.3046794","DOIUrl":null,"url":null,"abstract":"<p><p>Endoscopic images are used at various stages of rectal cancer treatment starting from cancer screening and diagnosis, during treatment to assess response and toxicity from treatments such as colitis, and at follow-up to detect new tumor or local regrowth. However, subjective assessment is highly variable and can underestimate the degree of response in some patients, subjecting them to unnecessary surgery, or overestimating response that places patients at risk of disease spread. Advances in deep learning have shown the ability to produce consistent and objective response assessments for endoscopic images. However, methods for detecting cancers, regrowth, and monitoring response during the entire course of patient treatment and follow-up are lacking. This is because automated diagnosis and rectal cancer response assessment require methods that are robust to inherent imaging illumination variations and confounding conditions (blood, scope, blurring) present in endoscopy images as well as changes to the normal lumen and tumor during treatment. Hence, a hierarchical shifted window (Swin) transformer was trained to distinguish rectal cancer from normal lumen using endoscopy images. Swin, as well as two convolutional (ResNet-50, WideResNet-50), and the vision transformer architectures, were trained and evaluated on follow-up longitudinal images to detect LR on in-distribution (ID) private datasets as well as on out-of-distribution (OOD) public colonoscopy datasets to detect pre/non-cancerous polyps. Color shifts were applied using optimal transport to simulate distribution shifts. Swin and ResNet models were similarly accurate in the ID dataset. Swin was more accurate than other methods (follow-up: 0.84, OOD: 0.83), even when subject to color shifts (follow-up: 0.83, OOD: 0.87), indicating the capability to provide robust performance for longitudinal cancer assessment.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"13406 ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12463470/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of SPIE--the International Society for Optical Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.3046794","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/4/11 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Endoscopic images are used at various stages of rectal cancer treatment starting from cancer screening and diagnosis, during treatment to assess response and toxicity from treatments such as colitis, and at follow-up to detect new tumor or local regrowth. However, subjective assessment is highly variable and can underestimate the degree of response in some patients, subjecting them to unnecessary surgery, or overestimating response that places patients at risk of disease spread. Advances in deep learning have shown the ability to produce consistent and objective response assessments for endoscopic images. However, methods for detecting cancers, regrowth, and monitoring response during the entire course of patient treatment and follow-up are lacking. This is because automated diagnosis and rectal cancer response assessment require methods that are robust to inherent imaging illumination variations and confounding conditions (blood, scope, blurring) present in endoscopy images as well as changes to the normal lumen and tumor during treatment. Hence, a hierarchical shifted window (Swin) transformer was trained to distinguish rectal cancer from normal lumen using endoscopy images. Swin, as well as two convolutional (ResNet-50, WideResNet-50), and the vision transformer architectures, were trained and evaluated on follow-up longitudinal images to detect LR on in-distribution (ID) private datasets as well as on out-of-distribution (OOD) public colonoscopy datasets to detect pre/non-cancerous polyps. Color shifts were applied using optimal transport to simulate distribution shifts. Swin and ResNet models were similarly accurate in the ID dataset. Swin was more accurate than other methods (follow-up: 0.84, OOD: 0.83), even when subject to color shifts (follow-up: 0.83, OOD: 0.87), indicating the capability to provide robust performance for longitudinal cancer assessment.

在基于内窥镜的纵向直肠癌评估中,Swin变压器对分布和概念漂移具有鲁棒性。
在直肠癌治疗的各个阶段,从癌症筛查和诊断开始,在治疗期间评估结肠炎等治疗的反应和毒性,以及在随访中发现新肿瘤或局部再生,都使用内镜图像。然而,主观评估是高度可变的,可能会低估一些患者的反应程度,使他们接受不必要的手术,或者高估反应,使患者面临疾病传播的风险。深度学习的进步已经显示出对内窥镜图像产生一致和客观的响应评估的能力。然而,在整个患者治疗和随访过程中,缺乏检测癌症、再生和监测反应的方法。这是因为自动诊断和直肠癌反应评估需要对内窥镜图像中存在的固有成像照明变化和混淆条件(血液、范围、模糊)以及治疗期间正常腔和肿瘤的变化具有鲁棒性的方法。因此,我们训练了一个分层移位窗口(Swin)转换器,通过内窥镜图像来区分直肠癌和正常腔。Swin以及两个卷积(ResNet-50, WideResNet-50)和视觉转换器架构,在后续纵向图像上进行训练和评估,以检测分布内(ID)私有数据集上的LR,以及分布外(OOD)公共结肠镜检查数据集上的LR,以检测癌前/非癌性息肉。颜色偏移采用最优运输来模拟分布偏移。Swin和ResNet模型在ID数据集中同样准确。Swin比其他方法更准确(随访:0.84,OOD: 0.83),即使受到颜色变化的影响(随访:0.83,OOD: 0.87),表明能够为纵向癌症评估提供可靠的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
0.50
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信