AVP-AP: Self-Supervised Automatic View Positioning in 3D Cardiac CT via Atlas Prompting

Xiaolin Fan;Yan Wang;Yingying Zhang;Mingkun Bao;Bosen Jia;Dong Lu;Yifan Gu;Jian Cheng;Haogang Zhu
{"title":"AVP-AP: Self-Supervised Automatic View Positioning in 3D Cardiac CT via Atlas Prompting","authors":"Xiaolin Fan;Yan Wang;Yingying Zhang;Mingkun Bao;Bosen Jia;Dong Lu;Yifan Gu;Jian Cheng;Haogang Zhu","doi":"10.1109/TMI.2025.3554785","DOIUrl":null,"url":null,"abstract":"Automatic view positioning is crucial for cardiac computed tomography (CT) examinations, including disease diagnosis and surgical planning. However, it is highly challenging due to individual variability and large 3D search space. Existing work needs labor-intensive and time-consuming manual annotations to train view-specific models, which are limited to predicting only a fixed set of planes. However, in real clinical scenarios, the challenge of positioning semantic 2D slices with any orientation into varying coordinate space in arbitrary 3D volume remains unsolved. We thus introduce a novel framework, AVP-AP, the first to use Atlas Prompting for self-supervised Automatic View Positioning in the 3D CT volume. Specifically, this paper first proposes an atlas prompting method, which generates a 3D canonical atlas and trains a network to map slices into their corresponding positions in the atlas space via a self-supervised manner. Then, guided by atlas prompts corresponding to the given query images in a reference CT, we identify the coarse positions of slices in the target CT volume using rigid transformation between the 3D atlas and target CT volume, effectively reducing the search space. Finally, we refine the coarse positions by maximizing the similarity between the predicted slices and the query images in the feature space of a given foundation model. Our framework is flexible and efficient compared to other methods, outperforming other methods by 19.8% average structural similarity (SSIM) in arbitrary view positioning and achieving 9% SSIM in two-chamber view compared to four radiologists. Meanwhile, experiments on a public dataset validate our framework’s generalizability.","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"44 7","pages":"2921-2932"},"PeriodicalIF":0.0000,"publicationDate":"2025-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on medical imaging","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10942463/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Automatic view positioning is crucial for cardiac computed tomography (CT) examinations, including disease diagnosis and surgical planning. However, it is highly challenging due to individual variability and large 3D search space. Existing work needs labor-intensive and time-consuming manual annotations to train view-specific models, which are limited to predicting only a fixed set of planes. However, in real clinical scenarios, the challenge of positioning semantic 2D slices with any orientation into varying coordinate space in arbitrary 3D volume remains unsolved. We thus introduce a novel framework, AVP-AP, the first to use Atlas Prompting for self-supervised Automatic View Positioning in the 3D CT volume. Specifically, this paper first proposes an atlas prompting method, which generates a 3D canonical atlas and trains a network to map slices into their corresponding positions in the atlas space via a self-supervised manner. Then, guided by atlas prompts corresponding to the given query images in a reference CT, we identify the coarse positions of slices in the target CT volume using rigid transformation between the 3D atlas and target CT volume, effectively reducing the search space. Finally, we refine the coarse positions by maximizing the similarity between the predicted slices and the query images in the feature space of a given foundation model. Our framework is flexible and efficient compared to other methods, outperforming other methods by 19.8% average structural similarity (SSIM) in arbitrary view positioning and achieving 9% SSIM in two-chamber view compared to four radiologists. Meanwhile, experiments on a public dataset validate our framework’s generalizability.
AVP-AP:基于Atlas提示的三维心脏CT自监督自动视图定位
自动视角定位对于心脏CT检查至关重要,包括疾病诊断和手术计划。然而,由于个体的可变性和大的3D搜索空间,这是极具挑战性的。现有的工作需要耗费大量劳动和时间的手工注释来训练特定于视图的模型,这些模型仅限于预测一组固定的平面。然而,在实际的临床场景中,将任意方向的语义二维切片定位到任意三维体的不同坐标空间的挑战仍然没有得到解决。因此,我们引入了一个新的框架,AVP-AP,这是第一个在3D CT体中使用Atlas提示进行自我监督自动视图定位的框架。具体而言,本文首先提出了一种地图集提示方法,该方法生成三维规范地图集,并训练网络以自监督的方式将切片映射到地图集空间中的相应位置。然后,在参考CT中给定查询图像对应的图集提示符的指导下,利用三维图集与目标CT体积之间的刚性变换,识别出目标CT体积中切片的粗糙位置,有效地缩小了搜索空间。最后,我们通过在给定基础模型的特征空间中最大化预测切片与查询图像之间的相似性来细化粗糙位置。与其他方法相比,我们的框架灵活高效,在任意视图定位中比其他方法高出19.8%的平均结构相似性(SSIM),在双腔视图中与四位放射科医生相比达到9%的SSIM。同时,在公共数据集上的实验验证了我们的框架的泛化性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信