阿联酋:多模态医学图像的通用解剖嵌入

IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Xiaoyu Bai , Fan Bai , Xiaofei Huo , Jia Ge , Jingjing Lu , Xianghua Ye , Minglei Shu , Ke Yan , Yong Xia
{"title":"阿联酋:多模态医学图像的通用解剖嵌入","authors":"Xiaoyu Bai ,&nbsp;Fan Bai ,&nbsp;Xiaofei Huo ,&nbsp;Jia Ge ,&nbsp;Jingjing Lu ,&nbsp;Xianghua Ye ,&nbsp;Minglei Shu ,&nbsp;Ke Yan ,&nbsp;Yong Xia","doi":"10.1016/j.media.2025.103562","DOIUrl":null,"url":null,"abstract":"<div><div>Identifying anatomical structures (<em>e.g.</em>, lesions or landmarks) is crucial for medical image analysis. Exemplar-based landmark detection methods are gaining attention as they allow the detection of arbitrary points during inference without needing annotated landmarks during training. These methods use self-supervised learning to create a discriminative voxel embedding and match corresponding landmarks via nearest-neighbor searches, showing promising results. However, current methods still face challenges in (1) differentiating voxels with similar appearance but different semantic meanings (<em>e.g.</em>, two adjacent structures without clear borders); (2) matching voxels with similar semantics but markedly different appearance (<em>e.g.</em>, the same vessel before and after contrast injection); and (3) cross-modality matching (<em>e.g.</em>, CT-MRI landmark-based registration). To overcome these challenges, we propose a Unified framework for learning Anatomical Embeddings (UAE). UAE is designed to learn appearance, semantic, and cross-modality anatomical embeddings. Specifically, UAE incorporates three key innovations: (1) semantic embedding learning with prototypical contrastive loss; (2) a fixed-point-based matching strategy; and (3) an iterative approach for cross-modality embedding learning. We thoroughly evaluated UAE across intra- and inter-modality tasks, including one-shot landmark detection, lesion tracking on longitudinal CT scans, and CT-MRI affine/rigid registration with varying fields of view. Our results suggest that UAE outperforms state-of-the-art methods, offering a robust and versatile approach for landmark-based medical image analysis tasks. Code and trained models are available at: <span><span>https://shorturl.at/bgsB3</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"103 ","pages":"Article 103562"},"PeriodicalIF":10.7000,"publicationDate":"2025-04-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"UAE: Universal Anatomical Embedding on multi-modality medical images\",\"authors\":\"Xiaoyu Bai ,&nbsp;Fan Bai ,&nbsp;Xiaofei Huo ,&nbsp;Jia Ge ,&nbsp;Jingjing Lu ,&nbsp;Xianghua Ye ,&nbsp;Minglei Shu ,&nbsp;Ke Yan ,&nbsp;Yong Xia\",\"doi\":\"10.1016/j.media.2025.103562\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Identifying anatomical structures (<em>e.g.</em>, lesions or landmarks) is crucial for medical image analysis. Exemplar-based landmark detection methods are gaining attention as they allow the detection of arbitrary points during inference without needing annotated landmarks during training. These methods use self-supervised learning to create a discriminative voxel embedding and match corresponding landmarks via nearest-neighbor searches, showing promising results. However, current methods still face challenges in (1) differentiating voxels with similar appearance but different semantic meanings (<em>e.g.</em>, two adjacent structures without clear borders); (2) matching voxels with similar semantics but markedly different appearance (<em>e.g.</em>, the same vessel before and after contrast injection); and (3) cross-modality matching (<em>e.g.</em>, CT-MRI landmark-based registration). To overcome these challenges, we propose a Unified framework for learning Anatomical Embeddings (UAE). UAE is designed to learn appearance, semantic, and cross-modality anatomical embeddings. Specifically, UAE incorporates three key innovations: (1) semantic embedding learning with prototypical contrastive loss; (2) a fixed-point-based matching strategy; and (3) an iterative approach for cross-modality embedding learning. We thoroughly evaluated UAE across intra- and inter-modality tasks, including one-shot landmark detection, lesion tracking on longitudinal CT scans, and CT-MRI affine/rigid registration with varying fields of view. Our results suggest that UAE outperforms state-of-the-art methods, offering a robust and versatile approach for landmark-based medical image analysis tasks. Code and trained models are available at: <span><span>https://shorturl.at/bgsB3</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":18328,\"journal\":{\"name\":\"Medical image analysis\",\"volume\":\"103 \",\"pages\":\"Article 103562\"},\"PeriodicalIF\":10.7000,\"publicationDate\":\"2025-04-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Medical image analysis\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1361841525001094\",\"RegionNum\":1,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical image analysis","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1361841525001094","RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

识别解剖结构(例如,病变或地标)对于医学图像分析至关重要。基于样本的地标检测方法越来越受到关注,因为它们允许在推理过程中检测任意点,而不需要在训练过程中标注地标。这些方法使用自监督学习来创建判别体素嵌入,并通过最近邻搜索匹配相应的地标,显示出令人满意的结果。然而,目前的方法仍然面临着以下挑战:(1)区分外观相似但语义不同的体素(例如,两个相邻的结构没有明确的边界);(2)匹配语义相似但外观明显不同的体素(例如,注入对比剂前后相同的血管);(3)交叉模态匹配(例如,基于CT-MRI地标的配准)。为了克服这些挑战,我们提出了一个统一的框架来学习解剖嵌入(UAE)。UAE旨在学习外观、语义和跨模态解剖嵌入。具体而言,阿联酋包含三个关键创新:(1)具有原型对比损失的语义嵌入学习;(2)基于定点的匹配策略;(3)跨模态嵌入学习的迭代方法。我们通过模态内和模态间的任务全面评估了UAE,包括一次性地标检测,纵向CT扫描的病变跟踪,以及不同视场的CT- mri仿射/刚性配准。我们的研究结果表明,阿联酋优于最先进的方法,为基于地标的医学图像分析任务提供了强大而通用的方法。代码和经过训练的模型可在https://shorturl.at/bgsB3上获得。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
UAE: Universal Anatomical Embedding on multi-modality medical images
Identifying anatomical structures (e.g., lesions or landmarks) is crucial for medical image analysis. Exemplar-based landmark detection methods are gaining attention as they allow the detection of arbitrary points during inference without needing annotated landmarks during training. These methods use self-supervised learning to create a discriminative voxel embedding and match corresponding landmarks via nearest-neighbor searches, showing promising results. However, current methods still face challenges in (1) differentiating voxels with similar appearance but different semantic meanings (e.g., two adjacent structures without clear borders); (2) matching voxels with similar semantics but markedly different appearance (e.g., the same vessel before and after contrast injection); and (3) cross-modality matching (e.g., CT-MRI landmark-based registration). To overcome these challenges, we propose a Unified framework for learning Anatomical Embeddings (UAE). UAE is designed to learn appearance, semantic, and cross-modality anatomical embeddings. Specifically, UAE incorporates three key innovations: (1) semantic embedding learning with prototypical contrastive loss; (2) a fixed-point-based matching strategy; and (3) an iterative approach for cross-modality embedding learning. We thoroughly evaluated UAE across intra- and inter-modality tasks, including one-shot landmark detection, lesion tracking on longitudinal CT scans, and CT-MRI affine/rigid registration with varying fields of view. Our results suggest that UAE outperforms state-of-the-art methods, offering a robust and versatile approach for landmark-based medical image analysis tasks. Code and trained models are available at: https://shorturl.at/bgsB3.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Medical image analysis
Medical image analysis 工程技术-工程:生物医学
CiteScore
22.10
自引率
6.40%
发文量
309
审稿时长
6.6 months
期刊介绍: Medical Image Analysis serves as a platform for sharing new research findings in the realm of medical and biological image analysis, with a focus on applications of computer vision, virtual reality, and robotics to biomedical imaging challenges. The journal prioritizes the publication of high-quality, original papers contributing to the fundamental science of processing, analyzing, and utilizing medical and biological images. It welcomes approaches utilizing biomedical image datasets across all spatial scales, from molecular/cellular imaging to tissue/organ imaging.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信