Farzad Beizaee , Gregory A. Lodygensky , Chris L. Adamson , Deanne K. Thompson , Jeanie L.Y. Cheong , Alicia J. Spittle , Peter J. Anderson , Christian Desrosiers , Jose Dolz
{"title":"Harmonizing flows: Leveraging normalizing flows for unsupervised and source-free MRI harmonization","authors":"Farzad Beizaee , Gregory A. Lodygensky , Chris L. Adamson , Deanne K. Thompson , Jeanie L.Y. Cheong , Alicia J. Spittle , Peter J. Anderson , Christian Desrosiers , Jose Dolz","doi":"10.1016/j.media.2025.103483","DOIUrl":"10.1016/j.media.2025.103483","url":null,"abstract":"<div><div>Lack of standardization and various intrinsic parameters for magnetic resonance (MR) image acquisition results in heterogeneous images across different sites and devices, which adversely affects the generalization of deep neural networks. To alleviate this issue, this work proposes a novel unsupervised harmonization framework that leverages normalizing flows to align MR images, thereby emulating the distribution of a source domain. The proposed strategy comprises three key steps. Initially, a normalizing flow network is trained to capture the distribution characteristics of the source domain. Then, we train a shallow harmonizer network to reconstruct images from the source domain via their augmented counterparts. Finally, during inference, the harmonizer network is updated to ensure that the output images conform to the learned source domain distribution, as modeled by the normalizing flow network. Our approach, which is unsupervised, source-free, and task-agnostic is assessed in the context of both adults and neonatal cross-domain brain MRI segmentation, as well as neonatal brain age estimation, demonstrating its generalizability across tasks and population demographics. The results underscore its superior performance compared to existing methodologies. The code is available at <span><span>https://github.com/farzad-bz/Harmonizing-Flows</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"101 ","pages":"Article 103483"},"PeriodicalIF":10.7,"publicationDate":"2025-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143353410","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Xiaosong Wang , Dequan Wang , Xiaoxiao Li , Jens Rittscher , Dimitris Metaxas , Shaoting Zhang
{"title":"Editorial for Special Issue on Foundation Models for Medical Image Analysis","authors":"Xiaosong Wang , Dequan Wang , Xiaoxiao Li , Jens Rittscher , Dimitris Metaxas , Shaoting Zhang","doi":"10.1016/j.media.2024.103389","DOIUrl":"10.1016/j.media.2024.103389","url":null,"abstract":"","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103389"},"PeriodicalIF":10.7,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142739884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Song Tang , Shaxu Yan , Xiaozhi Qi , Jianxin Gao , Mao Ye , Jianwei Zhang , Xiatian Zhu
{"title":"Few-shot medical image segmentation with high-fidelity prototypes","authors":"Song Tang , Shaxu Yan , Xiaozhi Qi , Jianxin Gao , Mao Ye , Jianwei Zhang , Xiatian Zhu","doi":"10.1016/j.media.2024.103412","DOIUrl":"10.1016/j.media.2024.103412","url":null,"abstract":"<div><div>Few-shot Semantic Segmentation (FSS) aims to adapt a pretrained model to new classes with as few as a single labeled training sample per class. Despite the prototype based approaches have achieved substantial success, existing models are limited to the imaging scenarios with considerably distinct objects and not highly complex background, e.g., natural images. This makes such models suboptimal for medical imaging with both conditions invalid. To address this problem, we propose a novel <strong>D</strong><em>etail</em> <strong>S</strong><em>elf-refined</em> <strong>P</strong><em>rototype</em> <strong>Net</strong><em>work</em> (<strong>DSPNet</strong>) to construct high-fidelity prototypes representing the object foreground and the background more comprehensively. Specifically, to construct global semantics while maintaining the captured detail semantics, we learn the foreground prototypes by modeling the multimodal structures with clustering and then fusing each in a channel-wise manner. Considering that the background often has no apparent semantic relation in the spatial dimensions, we integrate channel-specific structural information under sparse channel-aware regulation. Extensive experiments on three challenging medical image benchmarks show the superiority of DSPNet over previous state-of-the-art methods. The code and data are available at <span><span>https://github.com/tntek/DSPNet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103412"},"PeriodicalIF":10.7,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142780633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"COLLATOR: Consistent spatial–temporal longitudinal atlas construction via implicit neural representation","authors":"Lixuan Chen , Xuanyu Tian , Jiangjie Wu , Guoyan Lao , Yuyao Zhang , Hongjiang Wei","doi":"10.1016/j.media.2024.103396","DOIUrl":"10.1016/j.media.2024.103396","url":null,"abstract":"<div><div>Longitudinal brain atlases that present brain development trend along time, are essential tools for brain development studies. However, conventional methods construct these atlases by independently averaging brain images from different individuals at discrete time points. This approach could introduce temporal inconsistencies due to variations in ontogenetic trends among samples, potentially affecting accuracy of brain developmental characteristic analysis. In this paper, we propose an implicit neural representation (INR)-based framework to improve the temporal consistency in longitudinal atlases. We treat temporal inconsistency as a 4-dimensional (4D) image denoising task, where the data consists of 3D spatial information and 1D temporal progression. We formulate the longitudinal atlas as an implicit function of the spatial–temporal coordinates, allowing structural inconsistency over the time to be considered as 3D image noise along age. Inspired by recent self-supervised denoising methods (e.g. Noise2Noise), our approach learns the noise-free and temporally continuous implicit function from inconsistent longitudinal atlas data. Finally, the time-consistent longitudinal brain atlas can be reconstructed by evaluating the denoised 4D INR function at critical brain developing time points. We evaluate our approach on three longitudinal brain atlases of different MRI modalities, demonstrating that our method significantly improves temporal consistency while accurately preserving brain structures. Additionally, the continuous functions generated by our method enable the creation of 4D atlases with higher spatial and temporal resolution. Code: <span><span>https://github.com/maopaom/COLLATOR</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103396"},"PeriodicalIF":10.7,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142789979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Matthias Ivantsits , Leonid Goubergrits , Jan-Martin Kuhnigk , Markus Huellebrand , Jan Bruening , Tabea Kossen , Boris Pfahringer , Jens Schaller , Andreas Spuler , Titus Kuehne , Yizhuan Jia , Xuesong Li , Suprosanna Shit , Bjoern Menze , Ziyu Su , Jun Ma , Ziwei Nie , Kartik Jain , Yanfei Liu , Yi Lin , Anja Hennemuth
{"title":"Corrigendum to “Detection and analysis of cerebral aneurysms based on X-ray rotational angiography - the CADA 2020 challenge” [Medical Image Analysis, April 2022, Volume 77, 102333]","authors":"Matthias Ivantsits , Leonid Goubergrits , Jan-Martin Kuhnigk , Markus Huellebrand , Jan Bruening , Tabea Kossen , Boris Pfahringer , Jens Schaller , Andreas Spuler , Titus Kuehne , Yizhuan Jia , Xuesong Li , Suprosanna Shit , Bjoern Menze , Ziyu Su , Jun Ma , Ziwei Nie , Kartik Jain , Yanfei Liu , Yi Lin , Anja Hennemuth","doi":"10.1016/j.media.2024.103363","DOIUrl":"10.1016/j.media.2024.103363","url":null,"abstract":"","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103363"},"PeriodicalIF":10.7,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142400702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Qiang Ma , Kaili Liang , Liu Li , Saga Masui , Yourong Guo , Chiara Nosarti , Emma C. Robinson , Bernhard Kainz , Daniel Rueckert
{"title":"The Developing Human Connectome Project: A fast deep learning-based pipeline for neonatal cortical surface reconstruction","authors":"Qiang Ma , Kaili Liang , Liu Li , Saga Masui , Yourong Guo , Chiara Nosarti , Emma C. Robinson , Bernhard Kainz , Daniel Rueckert","doi":"10.1016/j.media.2024.103394","DOIUrl":"10.1016/j.media.2024.103394","url":null,"abstract":"<div><div>The Developing Human Connectome Project (dHCP) aims to explore developmental patterns of the human brain during the perinatal period. An automated processing pipeline has been developed to extract high-quality cortical surfaces from structural brain magnetic resonance (MR) images for the dHCP neonatal dataset. However, the current implementation of the pipeline requires more than 6.5 h to process a single MRI scan, making it expensive for large-scale neuroimaging studies. In this paper, we propose a fast deep learning (DL) based pipeline for dHCP neonatal cortical surface reconstruction, incorporating DL-based brain extraction, cortical surface reconstruction and spherical projection, as well as GPU-accelerated cortical surface inflation and cortical feature estimation. We introduce a multiscale deformation network to learn diffeomorphic cortical surface reconstruction end-to-end from T2-weighted brain MRI. A fast unsupervised spherical mapping approach is integrated to minimize metric distortions between cortical surfaces and projected spheres. The entire workflow of our DL-based dHCP pipeline completes within only 24 s on a modern GPU, which is nearly 1000 times faster than the original dHCP pipeline. The qualitative assessment demonstrates that for 82.5% of the test samples, the cortical surfaces reconstructed by our DL-based pipeline achieve superior (54.2%) or equal (28.3%) surface quality compared to the original dHCP pipeline.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103394"},"PeriodicalIF":10.7,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142780635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Rina Bao , Rebecca J. Weiss , Sara V. Bates , Ya’nan Song , Sheng He , Jingpeng Li , Alte Bjornerud , Randy L. Hirschtick , P. Ellen Grant , Yangming Ou
{"title":"PARADISE: Personalized and regional adaptation for HIE disease identification and segmentation","authors":"Rina Bao , Rebecca J. Weiss , Sara V. Bates , Ya’nan Song , Sheng He , Jingpeng Li , Alte Bjornerud , Randy L. Hirschtick , P. Ellen Grant , Yangming Ou","doi":"10.1016/j.media.2024.103419","DOIUrl":"10.1016/j.media.2024.103419","url":null,"abstract":"<div><div>Hypoxic ischemic encephalopathy (HIE) is a brain dysfunction occurring in approximately 1-5/1000 term-born neonates. Accurate segmentation of HIE lesions in brain MRI is crucial for prognosis and diagnosis but presents a unique challenge due to the diffuse and small nature of these abnormalities, which resulted in a substantial gap between the performance of machine learning-based segmentation methods and clinical expert annotations for HIE. To address this challenge, we introduce ParadiseNet, an algorithm specifically designed for HIE lesion segmentation. ParadiseNet incorporates global–local learning, progressive uncertainty learning, and self-evolution learning modules, all inspired by clinical interpretation of neonatal brain MRIs. These modules target issues such as unbalanced data distribution, boundary uncertainty, and imprecise lesion detection, respectively. Extensive experiments demonstrate that ParadiseNet significantly enhances small lesion detection (<span><math><mrow><mo><</mo><mn>1</mn><mtext>%</mtext></mrow></math></span>) accuracy in HIE, achieving an over 4% improvement in Dice, 6% improvement in NSD compared to U-Net and other general medical image segmentation algorithms.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"102 ","pages":"Article 103419"},"PeriodicalIF":10.7,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143705382","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yejee Shin , Geonhui Son , Dosik Hwang , Taejoon Eo
{"title":"Ensemble and low-frequency mixing with diffusion models for accelerated MRI reconstruction","authors":"Yejee Shin , Geonhui Son , Dosik Hwang , Taejoon Eo","doi":"10.1016/j.media.2025.103477","DOIUrl":"10.1016/j.media.2025.103477","url":null,"abstract":"<div><div>Magnetic resonance imaging (MRI) is an important imaging modality in medical diagnosis, providing comprehensive anatomical information with detailed tissue structures. However, the long scan time required to acquire high-quality MR images is a major challenge, especially in urgent clinical scenarios. Although diffusion models have achieved remarkable performance in accelerated MRI, there are several challenges. In particular, they struggle with the long inference time due to the high number of iterations in the reverse process of diffusion models. Additionally, they occasionally create artifacts or ‘hallucinate’ tissues that do not exist in the original anatomy. To address these problems, we propose ensemble and adaptive low-frequency mixing on the diffusion model, namely ELF-Diff for accelerated MRI. The proposed method consists of three key components in the reverse diffusion step: (1) optimization based on unified data consistency; (2) low-frequency mixing; and (3) aggregation of multiple perturbations of the predicted images for the ensemble in each step. We evaluate ELF-Diff on two MRI datasets, FastMRI and SKM-TEA. ELF-Diff surpasses other existing diffusion models for MRI reconstruction. Furthermore, extensive experiments, including a subtask of pathology detection, further demonstrate the superior anatomical precision of our method. ELF-Diff outperforms the existing state-of-the-art MRI reconstruction methods without being limited to specific undersampling patterns.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"101 ","pages":"Article 103477"},"PeriodicalIF":10.7,"publicationDate":"2025-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143296486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Xiangxin Wang , Zhan Wu , Yujia Zhou , Huazhong Shu , Jean-Louis Coatrieux , Qianjin Feng , Yang Chen
{"title":"Topology-oriented foreground focusing network for semi-supervised coronary artery segmentation","authors":"Xiangxin Wang , Zhan Wu , Yujia Zhou , Huazhong Shu , Jean-Louis Coatrieux , Qianjin Feng , Yang Chen","doi":"10.1016/j.media.2025.103465","DOIUrl":"10.1016/j.media.2025.103465","url":null,"abstract":"<div><div>Automatic coronary artery (CA) segmentation on coronary-computed tomography angiography (CCTA) images is critical for coronary-related disease diagnosis and pre-operative planning. However, such segmentation remains a challenging task due to the difficulty in maintaining the topological consistency of CA, interference from irrelevant tubular structures, and insufficient labeled data. In this study, we propose a novel semi-supervised topology-oriented foreground focusing network (TOFF-Net) to comprehensively address such challenges. Specifically, we first propose an explicit vascular connectivity preservation (VCP) loss to capture the topological information and effectively strengthen vascular connectivity. Then, we propose an irrelevant vessels removal (IVR) module, which aims to integrate local CA details and global CA distribution, thereby eliminating interference of irrelevant vessels. Moreover, we propose a foreground label migration and focusing (FLMF) module with Pioneer-Imitator learning as a semi-supervised strategy to exploit the unlabeled data. The FLMF can effectively guide the attention of TOFF-Net to the foreground. Extensive results on our in-house dataset and two public datasets demonstrate that our TOFF-Net achieves state-of-the-art CA segmentation performance with high topological consistency and few false-positive irrelevant tubular structures. The results also reveal that our TOFF-Net presents considerable potential for parsing other types of vessels.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"101 ","pages":"Article 103465"},"PeriodicalIF":10.7,"publicationDate":"2025-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143445555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Adeleh Bitarafan , Mohammad Mozafari , Mohammad Farid Azampour , Mahdieh Soleymani Baghshah , Nassir Navab , Azade Farshad
{"title":"Self-supervised 3D medical image segmentation by flow-guided mask propagation learning","authors":"Adeleh Bitarafan , Mohammad Mozafari , Mohammad Farid Azampour , Mahdieh Soleymani Baghshah , Nassir Navab , Azade Farshad","doi":"10.1016/j.media.2025.103478","DOIUrl":"10.1016/j.media.2025.103478","url":null,"abstract":"<div><div>Despite significant progress in 3D medical image segmentation using deep learning, manual annotation remains a labor-intensive bottleneck. Self-supervised mask propagation (SMP) methods have emerged to alleviate this challenge, allowing intra-volume segmentation with just a single slice annotation. However, the previous SMP methods often rely on 2D information and ignore volumetric contexts. While our previous work, called Vol2Flow, attempts to address this concern, it exhibits limitations, including not focusing enough on local (i.e., slice-pair) information, neglecting global information (i.e., volumetric contexts) in the objective function, and error accumulation during slice-to-slice reconstruction. This paper introduces Flow2Mask, a novel SMP method, developed to overcome the limitations of previous SMP approaches, particularly Vol2Flow. During training, Flow2Mask proposes the <em>Local-to-Global (L2G)</em> loss to learn inter-slice flow fields among all consecutive slices within a volume in an unsupervised manner. This dynamic loss is based on curriculum learning to gradually learn information within a volume from local to global contexts. Additionally, the <em>Inter-Slice Smoothness (ISS)</em> loss is introduced as a regularization term to encourage changes between the slices occur consistently and continuously. During inference, Flow2Mask leverages these 3D flow fields for inter-slice mask propagation in a 3D image, spreading annotation from a single annotated slice to the entire volume. Moreover, we propose an automatic strategy to select the most representative slice as initial annotation in the mask propagation process. Experimental evaluations on different abdominal datasets demonstrate that our proposed SMP method outperforms previous approaches and improves the overall mean DSC of Vol2Flow by <span><math><mrow><mo>+</mo><mn>2</mn><mo>.</mo><mn>1</mn><mtext>%</mtext></mrow></math></span>, <span><math><mrow><mo>+</mo><mn>8</mn><mo>.</mo><mn>2</mn><mtext>%</mtext></mrow></math></span>, and <span><math><mrow><mo>+</mo><mn>4</mn><mo>.</mo><mn>0</mn><mtext>%</mtext></mrow></math></span> for the Sliver, CHAOS, and 3D-IRCAD datasets, respectively. Furthermore, Flow2Mask even exhibits substantial improvements in weakly-supervised and self-supervised few-shot segmentation methods when applied as a mask completion tool. The code and model for Flow2Mask are available at <span><span>https://github.com/AdelehBitarafan/Flow2Mask</span><svg><path></path></svg></span>, providing a valuable contribution to the field of medical image segmentation.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"101 ","pages":"Article 103478"},"PeriodicalIF":10.7,"publicationDate":"2025-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143428900","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}