Xiaoyu Bai , Fan Bai , Xiaofei Huo , Jia Ge , Jingjing Lu , Xianghua Ye , Minglei Shu , Ke Yan , Yong Xia
{"title":"阿联酋:多模态医学图像的通用解剖嵌入","authors":"Xiaoyu Bai , Fan Bai , Xiaofei Huo , Jia Ge , Jingjing Lu , Xianghua Ye , Minglei Shu , Ke Yan , Yong Xia","doi":"10.1016/j.media.2025.103562","DOIUrl":null,"url":null,"abstract":"<div><div>Identifying anatomical structures (<em>e.g.</em>, lesions or landmarks) is crucial for medical image analysis. Exemplar-based landmark detection methods are gaining attention as they allow the detection of arbitrary points during inference without needing annotated landmarks during training. These methods use self-supervised learning to create a discriminative voxel embedding and match corresponding landmarks via nearest-neighbor searches, showing promising results. However, current methods still face challenges in (1) differentiating voxels with similar appearance but different semantic meanings (<em>e.g.</em>, two adjacent structures without clear borders); (2) matching voxels with similar semantics but markedly different appearance (<em>e.g.</em>, the same vessel before and after contrast injection); and (3) cross-modality matching (<em>e.g.</em>, CT-MRI landmark-based registration). To overcome these challenges, we propose a Unified framework for learning Anatomical Embeddings (UAE). UAE is designed to learn appearance, semantic, and cross-modality anatomical embeddings. Specifically, UAE incorporates three key innovations: (1) semantic embedding learning with prototypical contrastive loss; (2) a fixed-point-based matching strategy; and (3) an iterative approach for cross-modality embedding learning. We thoroughly evaluated UAE across intra- and inter-modality tasks, including one-shot landmark detection, lesion tracking on longitudinal CT scans, and CT-MRI affine/rigid registration with varying fields of view. Our results suggest that UAE outperforms state-of-the-art methods, offering a robust and versatile approach for landmark-based medical image analysis tasks. Code and trained models are available at: <span><span>https://shorturl.at/bgsB3</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"103 ","pages":"Article 103562"},"PeriodicalIF":10.7000,"publicationDate":"2025-04-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"UAE: Universal Anatomical Embedding on multi-modality medical images\",\"authors\":\"Xiaoyu Bai , Fan Bai , Xiaofei Huo , Jia Ge , Jingjing Lu , Xianghua Ye , Minglei Shu , Ke Yan , Yong Xia\",\"doi\":\"10.1016/j.media.2025.103562\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Identifying anatomical structures (<em>e.g.</em>, lesions or landmarks) is crucial for medical image analysis. Exemplar-based landmark detection methods are gaining attention as they allow the detection of arbitrary points during inference without needing annotated landmarks during training. These methods use self-supervised learning to create a discriminative voxel embedding and match corresponding landmarks via nearest-neighbor searches, showing promising results. However, current methods still face challenges in (1) differentiating voxels with similar appearance but different semantic meanings (<em>e.g.</em>, two adjacent structures without clear borders); (2) matching voxels with similar semantics but markedly different appearance (<em>e.g.</em>, the same vessel before and after contrast injection); and (3) cross-modality matching (<em>e.g.</em>, CT-MRI landmark-based registration). To overcome these challenges, we propose a Unified framework for learning Anatomical Embeddings (UAE). UAE is designed to learn appearance, semantic, and cross-modality anatomical embeddings. Specifically, UAE incorporates three key innovations: (1) semantic embedding learning with prototypical contrastive loss; (2) a fixed-point-based matching strategy; and (3) an iterative approach for cross-modality embedding learning. We thoroughly evaluated UAE across intra- and inter-modality tasks, including one-shot landmark detection, lesion tracking on longitudinal CT scans, and CT-MRI affine/rigid registration with varying fields of view. Our results suggest that UAE outperforms state-of-the-art methods, offering a robust and versatile approach for landmark-based medical image analysis tasks. Code and trained models are available at: <span><span>https://shorturl.at/bgsB3</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":18328,\"journal\":{\"name\":\"Medical image analysis\",\"volume\":\"103 \",\"pages\":\"Article 103562\"},\"PeriodicalIF\":10.7000,\"publicationDate\":\"2025-04-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Medical image analysis\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1361841525001094\",\"RegionNum\":1,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical image analysis","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1361841525001094","RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
UAE: Universal Anatomical Embedding on multi-modality medical images
Identifying anatomical structures (e.g., lesions or landmarks) is crucial for medical image analysis. Exemplar-based landmark detection methods are gaining attention as they allow the detection of arbitrary points during inference without needing annotated landmarks during training. These methods use self-supervised learning to create a discriminative voxel embedding and match corresponding landmarks via nearest-neighbor searches, showing promising results. However, current methods still face challenges in (1) differentiating voxels with similar appearance but different semantic meanings (e.g., two adjacent structures without clear borders); (2) matching voxels with similar semantics but markedly different appearance (e.g., the same vessel before and after contrast injection); and (3) cross-modality matching (e.g., CT-MRI landmark-based registration). To overcome these challenges, we propose a Unified framework for learning Anatomical Embeddings (UAE). UAE is designed to learn appearance, semantic, and cross-modality anatomical embeddings. Specifically, UAE incorporates three key innovations: (1) semantic embedding learning with prototypical contrastive loss; (2) a fixed-point-based matching strategy; and (3) an iterative approach for cross-modality embedding learning. We thoroughly evaluated UAE across intra- and inter-modality tasks, including one-shot landmark detection, lesion tracking on longitudinal CT scans, and CT-MRI affine/rigid registration with varying fields of view. Our results suggest that UAE outperforms state-of-the-art methods, offering a robust and versatile approach for landmark-based medical image analysis tasks. Code and trained models are available at: https://shorturl.at/bgsB3.
期刊介绍:
Medical Image Analysis serves as a platform for sharing new research findings in the realm of medical and biological image analysis, with a focus on applications of computer vision, virtual reality, and robotics to biomedical imaging challenges. The journal prioritizes the publication of high-quality, original papers contributing to the fundamental science of processing, analyzing, and utilizing medical and biological images. It welcomes approaches utilizing biomedical image datasets across all spatial scales, from molecular/cellular imaging to tissue/organ imaging.