从超声波数据中检测前列腺癌的图像变换器基准测试

Mohamed Harmanani, P. Wilson, F. Fooladgar, A. Jamzad, Mahdi Gilany, Minh Nguyen Nhat To, B. Wodlinger, P. Abolmaesumi, P. Mousavi
{"title":"从超声波数据中检测前列腺癌的图像变换器基准测试","authors":"Mohamed Harmanani, P. Wilson, F. Fooladgar, A. Jamzad, Mahdi Gilany, Minh Nguyen Nhat To, B. Wodlinger, P. Abolmaesumi, P. Mousavi","doi":"10.1117/12.3006049","DOIUrl":null,"url":null,"abstract":"PURPOSE: Deep learning methods for classifying prostate cancer (PCa) in ultrasound images typically employ convolutional networks (CNNs) to detect cancer in small regions of interest (ROI) along a needle trace region. However, this approach suffers from weak labelling, since the ground-truth histopathology labels do not describe the properties of individual ROIs. Recently, multi-scale approaches have sought to mitigate this issue by combining the context awareness of transformers with a CNN feature extractor to detect cancer from multiple ROIs using multiple-instance learning (MIL). In this work, we present a detailed study of several image transformer architectures for both ROI-scale and multi-scale classification, and a comparison of the performance of CNNs and transformers for ultrasound-based prostate cancer classification. We also design a novel multi-objective learning strategy that combines both ROI and core predictions to further mitigate label noise. METHODS: We evaluate 3 image transformers on ROI-scale cancer classification, then use the strongest model to tune a multi-scale classifier with MIL. We train our MIL models using our novel multi-objective learning strategy and compare our results to existing baselines. RESULTS: We find that for both ROI-scale and multi-scale PCa detection, image transformer backbones lag behind their CNN counterparts. This deficit in performance is even more noticeable for larger models. When using multi-objective learning, we can improve performance of MIL, with a 77.9% AUROC, a sensitivity of 75.9%, and a specificity of 66.3%. CONCLUSION: Convolutional networks are better suited for modelling sparse datasets of prostate ultrasounds, producing more robust features than transformers in PCa detection. Multi-scale methods remain the best architecture for this task, with multi-objective learning presenting an effective way to improve performance.","PeriodicalId":517504,"journal":{"name":"Medical Imaging 2024: Image-Guided Procedures, Robotic Interventions, and Modeling","volume":"48 15","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Benchmarking image transformers for prostate cancer detection from ultrasound data\",\"authors\":\"Mohamed Harmanani, P. Wilson, F. Fooladgar, A. Jamzad, Mahdi Gilany, Minh Nguyen Nhat To, B. Wodlinger, P. Abolmaesumi, P. Mousavi\",\"doi\":\"10.1117/12.3006049\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"PURPOSE: Deep learning methods for classifying prostate cancer (PCa) in ultrasound images typically employ convolutional networks (CNNs) to detect cancer in small regions of interest (ROI) along a needle trace region. However, this approach suffers from weak labelling, since the ground-truth histopathology labels do not describe the properties of individual ROIs. Recently, multi-scale approaches have sought to mitigate this issue by combining the context awareness of transformers with a CNN feature extractor to detect cancer from multiple ROIs using multiple-instance learning (MIL). In this work, we present a detailed study of several image transformer architectures for both ROI-scale and multi-scale classification, and a comparison of the performance of CNNs and transformers for ultrasound-based prostate cancer classification. We also design a novel multi-objective learning strategy that combines both ROI and core predictions to further mitigate label noise. METHODS: We evaluate 3 image transformers on ROI-scale cancer classification, then use the strongest model to tune a multi-scale classifier with MIL. We train our MIL models using our novel multi-objective learning strategy and compare our results to existing baselines. RESULTS: We find that for both ROI-scale and multi-scale PCa detection, image transformer backbones lag behind their CNN counterparts. This deficit in performance is even more noticeable for larger models. When using multi-objective learning, we can improve performance of MIL, with a 77.9% AUROC, a sensitivity of 75.9%, and a specificity of 66.3%. CONCLUSION: Convolutional networks are better suited for modelling sparse datasets of prostate ultrasounds, producing more robust features than transformers in PCa detection. Multi-scale methods remain the best architecture for this task, with multi-objective learning presenting an effective way to improve performance.\",\"PeriodicalId\":517504,\"journal\":{\"name\":\"Medical Imaging 2024: Image-Guided Procedures, Robotic Interventions, and Modeling\",\"volume\":\"48 15\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-03-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Medical Imaging 2024: Image-Guided Procedures, Robotic Interventions, and Modeling\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1117/12.3006049\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical Imaging 2024: Image-Guided Procedures, Robotic Interventions, and Modeling","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.3006049","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

目的:用于对超声图像中的前列腺癌(PCa)进行分类的深度学习方法通常采用卷积网络(CNN)来检测沿针迹区域的小感兴趣区(ROI)中的癌症。然而,这种方法存在标记能力弱的问题,因为地面真实组织病理学标签无法描述单个 ROI 的属性。最近,多尺度方法试图缓解这一问题,将变换器的上下文感知与 CNN 特征提取器相结合,利用多实例学习(MIL)从多个 ROI 检测癌症。在这项工作中,我们详细研究了用于 ROI 规模和多规模分类的几种图像变换器架构,并比较了 CNN 和变换器在基于超声波的前列腺癌分类中的性能。我们还设计了一种新颖的多目标学习策略,将 ROI 和核心预测结合起来,以进一步减轻标签噪声。方法:我们对 3 种图像变换器进行了 ROI 尺度癌症分类评估,然后使用最强的模型来调整 MIL 的多尺度分类器。我们使用新颖的多目标学习策略训练 MIL 模型,并将结果与现有基线进行比较。结果:我们发现,无论是 ROI 尺度还是多尺度 PCa 检测,图像变换器骨干都落后于 CNN 骨干。对于较大的模型,这种性能上的不足更为明显。在使用多目标学习时,我们可以提高 MIL 的性能,其 AUROC 为 77.9%,灵敏度为 75.9%,特异度为 66.3%。结论:卷积网络更适合对稀疏的前列腺超声波数据集建模,在 PCa 检测中能产生比变换器更强大的特征。多尺度方法仍然是这项任务的最佳架构,多目标学习是提高性能的有效方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Benchmarking image transformers for prostate cancer detection from ultrasound data
PURPOSE: Deep learning methods for classifying prostate cancer (PCa) in ultrasound images typically employ convolutional networks (CNNs) to detect cancer in small regions of interest (ROI) along a needle trace region. However, this approach suffers from weak labelling, since the ground-truth histopathology labels do not describe the properties of individual ROIs. Recently, multi-scale approaches have sought to mitigate this issue by combining the context awareness of transformers with a CNN feature extractor to detect cancer from multiple ROIs using multiple-instance learning (MIL). In this work, we present a detailed study of several image transformer architectures for both ROI-scale and multi-scale classification, and a comparison of the performance of CNNs and transformers for ultrasound-based prostate cancer classification. We also design a novel multi-objective learning strategy that combines both ROI and core predictions to further mitigate label noise. METHODS: We evaluate 3 image transformers on ROI-scale cancer classification, then use the strongest model to tune a multi-scale classifier with MIL. We train our MIL models using our novel multi-objective learning strategy and compare our results to existing baselines. RESULTS: We find that for both ROI-scale and multi-scale PCa detection, image transformer backbones lag behind their CNN counterparts. This deficit in performance is even more noticeable for larger models. When using multi-objective learning, we can improve performance of MIL, with a 77.9% AUROC, a sensitivity of 75.9%, and a specificity of 66.3%. CONCLUSION: Convolutional networks are better suited for modelling sparse datasets of prostate ultrasounds, producing more robust features than transformers in PCa detection. Multi-scale methods remain the best architecture for this task, with multi-objective learning presenting an effective way to improve performance.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信