Autodelineation of Treatment Target Volume for Radiation Therapy Using Large Language Model-Aided Multimodal Learning.

IF 6.4 1区 医学 Q1 ONCOLOGY
Praveenbalaji Rajendran, Yizheng Chen, Liang Qiu, Thomas Niedermayr, Wu Liu, Mark Buyyounouski, Hilary Bagshaw, Bin Han, Yong Yang, Nataliya Kovalchuk, Xuejun Gu, Steven Hancock, Lei Xing, Xianjin Dai
{"title":"Autodelineation of Treatment Target Volume for Radiation Therapy Using Large Language Model-Aided Multimodal Learning.","authors":"Praveenbalaji Rajendran, Yizheng Chen, Liang Qiu, Thomas Niedermayr, Wu Liu, Mark Buyyounouski, Hilary Bagshaw, Bin Han, Yong Yang, Nataliya Kovalchuk, Xuejun Gu, Steven Hancock, Lei Xing, Xianjin Dai","doi":"10.1016/j.ijrobp.2024.07.2149","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>Artificial intelligence-aided methods have made significant progress in the auto-delineation of normal tissues. However, these approaches struggle with the auto-contouring of radiation therapy target volume. Our goal was to model the delineation of target volume as a clinical decision-making problem, resolved by leveraging large language model-aided multimodal learning approaches.</p><p><strong>Methods and materials: </strong>A vision-language model, termed Medformer, has been developed, employing the hierarchical vision transformer as its backbone and incorporating large language models to extract text-rich features. The contextually embedded linguistic features are seamlessly integrated into visual features for language-aware visual encoding through the visual language attention module. Metrics, including Dice similarity coefficient (DSC), intersection over union (IOU), and 95th percentile Hausdorff distance (HD95), were used to quantitatively evaluate the performance of our model. The evaluation was conducted on an in-house prostate cancer data set and a public oropharyngeal carcinoma data set, totaling 668 subjects.</p><p><strong>Results: </strong>Our Medformer achieved a DSC of 0.81 ± 0.10 versus 0.72 ± 0.10, IOU of 0.73 ± 0.12 versus 0.65 ± 0.09, and HD95 of 9.86 ± 9.77 mm versus 19.13 ± 12.96 mm for delineation of gross tumor volume on the prostate cancer dataset. Similarly, on the oropharyngeal carcinoma dataset, it achieved a DSC of 0.77 ± 0.11 versus 0.72 ± 0.09, IOU of 0.70 ± 0.09 versus 0.65 ± 0.07, and HD95 of 7.52 ± 4.8 mm versus 13.63 ± 7.13 mm, representing significant improvements (P < 0.05). For delineating the clinical target volume, Medformer achieved a DSC of 0.91 ± 0.04, IOU of 0.85 ± 0.05, and HD95 of 2.98 ± 1.60 mm, comparable with other state-of-the-art algorithms.</p><p><strong>Conclusions: </strong>Auto-delineation of the treatment target based on multimodal learning outperforms conventional approaches that rely purely on visual features. Our method could be adopted into routine practice to rapidly contour clinical target volume/gross tumor volume.</p>","PeriodicalId":14215,"journal":{"name":"International Journal of Radiation Oncology Biology Physics","volume":" ","pages":"230-240"},"PeriodicalIF":6.4000,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Radiation Oncology Biology Physics","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1016/j.ijrobp.2024.07.2149","RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/8/6 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"ONCOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose: Artificial intelligence-aided methods have made significant progress in the auto-delineation of normal tissues. However, these approaches struggle with the auto-contouring of radiation therapy target volume. Our goal was to model the delineation of target volume as a clinical decision-making problem, resolved by leveraging large language model-aided multimodal learning approaches.

Methods and materials: A vision-language model, termed Medformer, has been developed, employing the hierarchical vision transformer as its backbone and incorporating large language models to extract text-rich features. The contextually embedded linguistic features are seamlessly integrated into visual features for language-aware visual encoding through the visual language attention module. Metrics, including Dice similarity coefficient (DSC), intersection over union (IOU), and 95th percentile Hausdorff distance (HD95), were used to quantitatively evaluate the performance of our model. The evaluation was conducted on an in-house prostate cancer data set and a public oropharyngeal carcinoma data set, totaling 668 subjects.

Results: Our Medformer achieved a DSC of 0.81 ± 0.10 versus 0.72 ± 0.10, IOU of 0.73 ± 0.12 versus 0.65 ± 0.09, and HD95 of 9.86 ± 9.77 mm versus 19.13 ± 12.96 mm for delineation of gross tumor volume on the prostate cancer dataset. Similarly, on the oropharyngeal carcinoma dataset, it achieved a DSC of 0.77 ± 0.11 versus 0.72 ± 0.09, IOU of 0.70 ± 0.09 versus 0.65 ± 0.07, and HD95 of 7.52 ± 4.8 mm versus 13.63 ± 7.13 mm, representing significant improvements (P < 0.05). For delineating the clinical target volume, Medformer achieved a DSC of 0.91 ± 0.04, IOU of 0.85 ± 0.05, and HD95 of 2.98 ± 1.60 mm, comparable with other state-of-the-art algorithms.

Conclusions: Auto-delineation of the treatment target based on multimodal learning outperforms conventional approaches that rely purely on visual features. Our method could be adopted into routine practice to rapidly contour clinical target volume/gross tumor volume.

利用大语言模型辅助多模态学习,自动划分放射治疗的治疗目标体积。
目的:人工智能(AI)辅助方法在正常组织的自动划线方面取得了重大进展。然而,这些方法在放疗靶区的自动轮廓划分方面却举步维艰。我们的目标是将靶区划分作为一个临床决策问题进行建模,并利用大型语言模型辅助多模态学习方法加以解决:我们开发了一种名为 Medformer 的视觉语言模型,以分层视觉转换器为骨干,并结合大型语言模型来提取丰富的文本特征。通过视觉语言注意模块,将上下文嵌入的语言特征无缝集成到视觉特征中,进行语言感知视觉编码。包括骰子相似系数(DSC)、交集大于联合(IOU)和第 95 百分位数豪斯多夫距离(HD95)在内的指标被用来定量评估我们模型的性能。评估是在一个内部前列腺癌数据集和一个公共口咽癌(OPC)数据集上进行的,共有 668 个受试者:在前列腺癌数据集上,我们的Medformer划定肿瘤总体积(GTV)的DSC为0.81±0.10比0.72±0.10,IOU为0.73±0.12比0.65±0.09,HD95为9.86±9.77毫米比19.13±12.96毫米。同样,在 OPC 数据集上,其 DSC 为 0.77 ± 0.11 对 0.72 ± 0.09,IOU 为 0.70 ± 0.09 对 0.65 ± 0.07,HD95 为 7.52 ± 4.8 mm 对 13.63 ± 7.13 mm,均有显著改善(p < 0.05)。在划定临床靶体积(CTV)方面,Medformer 的 DSC 为 0.91 ± 0.04,IOU 为 0.85 ± 0.05,HD95 为 2.98 ± 1.60 毫米,与其他最先进的算法相当:基于多模态学习的治疗目标自动划线方法优于纯粹依赖视觉特征的传统方法。我们的方法可用于常规实践,以快速勾画CTV/GTV轮廓。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
11.00
自引率
7.10%
发文量
2538
审稿时长
6.6 weeks
期刊介绍: International Journal of Radiation Oncology • Biology • Physics (IJROBP), known in the field as the Red Journal, publishes original laboratory and clinical investigations related to radiation oncology, radiation biology, medical physics, and both education and health policy as it relates to the field. This journal has a particular interest in original contributions of the following types: prospective clinical trials, outcomes research, and large database interrogation. In addition, it seeks reports of high-impact innovations in single or combined modality treatment, tumor sensitization, normal tissue protection (including both precision avoidance and pharmacologic means), brachytherapy, particle irradiation, and cancer imaging. Technical advances related to dosimetry and conformal radiation treatment planning are of interest, as are basic science studies investigating tumor physiology and the molecular biology underlying cancer and normal tissue radiation response.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信