Ching-Wei Wang , Tzu-Chien Liu , Po-Jen Lai , Hikam Muzakky , Yu-Chi Wang , Mu-Hsien Yu , Chia-Hua Wu , Tai-Kuang Chao
{"title":"Ensemble transformer-based multiple instance learning to predict pathological subtypes and tumor mutational burden from histopathological whole slide images of endometrial and colorectal cancer","authors":"Ching-Wei Wang , Tzu-Chien Liu , Po-Jen Lai , Hikam Muzakky , Yu-Chi Wang , Mu-Hsien Yu , Chia-Hua Wu , Tai-Kuang Chao","doi":"10.1016/j.media.2024.103372","DOIUrl":"10.1016/j.media.2024.103372","url":null,"abstract":"<div><div>In endometrial cancer (EC) and colorectal cancer (CRC), in addition to microsatellite instability, tumor mutational burden (TMB) has gradually gained attention as a genomic biomarker that can be used clinically to determine which patients may benefit from immune checkpoint inhibitors. High TMB is characterized by a large number of mutated genes, which encode aberrant tumor neoantigens, and implies a better response to immunotherapy. Hence, a part of EC and CRC patients associated with high TMB may have higher chances to receive immunotherapy. TMB measurement was mainly evaluated by whole-exome sequencing or next-generation sequencing, which was costly and difficult to be widely applied in all clinical cases. Therefore, an effective, efficient, low-cost and easily accessible tool is urgently needed to distinguish the TMB status of EC and CRC patients. In this study, we present a deep learning framework, namely Ensemble Transformer-based Multiple Instance Learning with Self-Supervised Learning Vision Transformer feature encoder (ETMIL-SSLViT), to predict pathological subtype and TMB status directly from the H&E stained whole slide images (WSIs) in EC and CRC patients, which is helpful for both pathological classification and cancer treatment planning. Our framework was evaluated on two different cancer cohorts, including an EC cohort with 918 histopathology WSIs from 529 patients and a CRC cohort with 1495 WSIs from 594 patients from The Cancer Genome Atlas. The experimental results show that the proposed methods achieved excellent performance and outperforming seven state-of-the-art (SOTA) methods in cancer subtype classification and TMB prediction on both cancer datasets. Fisher’s exact test further validated that the associations between the predictions of the proposed models and the actual cancer subtype or TMB status are both extremely strong (<span><math><mrow><mi>p</mi><mo><</mo><mn>0</mn><mo>.</mo><mn>001</mn></mrow></math></span>). These promising findings show the potential of our proposed methods to guide personalized treatment decisions by accurately predicting the EC and CRC subtype and the TMB status for effective immunotherapy planning for EC and CRC patients.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"99 ","pages":"Article 103372"},"PeriodicalIF":10.7,"publicationDate":"2024-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142503533","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Julia Camps , Zhinuo Jenny Wang , Ruben Doste , Lucas Arantes Berg , Maxx Holmes , Brodie Lawson , Jakub Tomek , Kevin Burrage , Alfonso Bueno-Orovio , Blanca Rodriguez
{"title":"Harnessing 12-lead ECG and MRI data to personalise repolarisation profiles in cardiac digital twin models for enhanced virtual drug testing","authors":"Julia Camps , Zhinuo Jenny Wang , Ruben Doste , Lucas Arantes Berg , Maxx Holmes , Brodie Lawson , Jakub Tomek , Kevin Burrage , Alfonso Bueno-Orovio , Blanca Rodriguez","doi":"10.1016/j.media.2024.103361","DOIUrl":"10.1016/j.media.2024.103361","url":null,"abstract":"<div><div>Cardiac digital twins are computational tools capturing key functional and anatomical characteristics of patient hearts for investigating disease phenotypes and predicting responses to therapy. When paired with large-scale computational resources and large clinical datasets, digital twin technology can enable virtual clinical trials on virtual cohorts to fast-track therapy development. Here, we present an open-source automated pipeline for personalising ventricular electrophysiological function based on routinely acquired magnetic resonance imaging (MRI) data and the standard 12-lead electrocardiogram (ECG).</div><div>Using MRI-based anatomical models, a sequential Monte-Carlo approximate Bayesian computational inference method is extended to infer electrical activation and repolarisation characteristics from the ECG. Fast simulations are conducted with a reaction-Eikonal model, including the Purkinje network and biophysically-detailed subcellular ionic current dynamics for repolarisation. For each patient, parameter uncertainty is represented by inferring an envelope of plausible ventricular models rather than a single one, which means that parameter uncertainty can be propagated to therapy evaluation. Furthermore, we have developed techniques for translating from reaction-Eikonal to monodomain simulations, which allows more realistic simulations of cardiac electrophysiology. The pipeline is demonstrated in three healthy subjects, where our inferred pseudo-diffusion reaction-Eikonal models reproduced the patient's ECG with a median Pearson's correlation coefficient of 0.9, and then translated to monodomain simulations with a median correlation coefficient of 0.84 across all subjects. We then demonstrate our digital twins for virtual evaluation of Dofetilide with uncertainty quantification. These evaluations using our cardiac digital twins reproduced dose-dependent QTc and T peak to T end prolongations that are in keeping with large population drug response data.</div><div>The methodologies for cardiac digital twinning presented here are a step towards personalised virtual therapy testing and can be scaled to generate virtual populations for clinical trials to fast-track therapy evaluation. The tools developed for this paper are open-source, documented, and made publicly available.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103361"},"PeriodicalIF":10.7,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142744160","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Fan Wang , Zhilin Zou , Nicole Sakla , Luke Partyka , Nil Rawal , Gagandeep Singh , Wei Zhao , Haibin Ling , Chuan Huang , Prateek Prasanna , Chao Chen
{"title":"TopoTxR: A topology-guided deep convolutional network for breast parenchyma learning on DCE-MRIs","authors":"Fan Wang , Zhilin Zou , Nicole Sakla , Luke Partyka , Nil Rawal , Gagandeep Singh , Wei Zhao , Haibin Ling , Chuan Huang , Prateek Prasanna , Chao Chen","doi":"10.1016/j.media.2024.103373","DOIUrl":"10.1016/j.media.2024.103373","url":null,"abstract":"<div><div>Characterization of breast parenchyma in dynamic contrast-enhanced magnetic resonance imaging (DCE-MRI) is a challenging task owing to the complexity of underlying tissue structures. Existing quantitative approaches, like radiomics and deep learning models, lack explicit quantification of intricate and subtle parenchymal structures, including fibroglandular tissue. To address this, we propose a novel topological approach that explicitly extracts multi-scale topological structures to better approximate breast parenchymal structures, and then incorporates these structures into a deep-learning-based prediction model via an attention mechanism. Our topology-informed deep learning model, <em>TopoTxR</em>, leverages topology to provide enhanced insights into tissues critical for disease pathophysiology and treatment response. We empirically validate <em>TopoTxR</em> using the VICTRE phantom breast dataset, showing that the topological structures extracted by our model effectively approximate the breast parenchymal structures. We further demonstrate <em>TopoTxR</em>’s efficacy in predicting response to neoadjuvant chemotherapy. Our qualitative and quantitative analyses suggest differential topological behavior of breast tissue in treatment-naïve imaging, in patients who respond favorably to therapy as achieving pathological complete response (pCR) versus those who do not. In a comparative analysis with several baselines on the publicly available I-SPY 1 dataset (N = 161, including 47 patients with pCR and 114 without) and the Rutgers proprietary dataset (N = 120, with 69 patients achieving pCR and 51 not), <em>TopoTxR</em> demonstrates a notable improvement, achieving a 2.6% increase in accuracy and a 4.6% enhancement in AUC compared to the state-of-the-art method.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"99 ","pages":"Article 103373"},"PeriodicalIF":10.7,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142503536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mojtaba Lashgari , Zheyi Yang , Miguel O. Bernabeu , Jing-Rebecca Li , Alejandro F. Frangi
{"title":"SpinDoctor-IVIM: A virtual imaging framework for intravoxel incoherent motion MRI","authors":"Mojtaba Lashgari , Zheyi Yang , Miguel O. Bernabeu , Jing-Rebecca Li , Alejandro F. Frangi","doi":"10.1016/j.media.2024.103369","DOIUrl":"10.1016/j.media.2024.103369","url":null,"abstract":"<div><div>Intravoxel incoherent motion (IVIM) imaging is increasingly recognised as an important tool in clinical MRI, where tissue perfusion and diffusion information can aid disease diagnosis, monitoring of patient recovery, and treatment outcome assessment. Currently, the discovery of biomarkers based on IVIM imaging, similar to other medical imaging modalities, is dependent on long preclinical and clinical validation pathways to link observable markers derived from images with the underlying pathophysiological mechanisms. To speed up this process, virtual IVIM imaging is proposed. This approach provides an efficient virtual imaging tool to design, evaluate, and optimise novel approaches for IVIM imaging. In this work, virtual IVIM imaging is developed through a new finite element solver, SpinDoctor-IVIM, which extends SpinDoctor, a diffusion MRI simulation toolbox. SpinDoctor-IVIM simulates IVIM imaging signals by solving the generalised Bloch–Torrey partial differential equation. The input velocity to SpinDoctor-IVIM is computed using HemeLB, an established Lattice Boltzmann blood flow simulator. Contrary to previous approaches, SpinDoctor-IVIM accounts for volumetric microvasculature during blood flow simulations, incorporates diffusion phenomena in the intravascular space, and accounts for the permeability between the intravascular and extravascular spaces. The above-mentioned features of the proposed framework are illustrated with simulations on a realistic microvasculature model.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"99 ","pages":"Article 103369"},"PeriodicalIF":10.7,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142503535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Wenhui Lei , Wei Xu , Kang Li , Xiaofan Zhang , Shaoting Zhang
{"title":"MedLSAM: Localize and segment anything model for 3D CT images","authors":"Wenhui Lei , Wei Xu , Kang Li , Xiaofan Zhang , Shaoting Zhang","doi":"10.1016/j.media.2024.103370","DOIUrl":"10.1016/j.media.2024.103370","url":null,"abstract":"<div><div>Recent advancements in foundation models have shown significant potential in medical image analysis. However, there is still a gap in models specifically designed for medical image localization. To address this, we introduce MedLAM, a 3D medical foundation localization model that accurately identifies any anatomical part within the body using only a few template scans. MedLAM employs two self-supervision tasks: unified anatomical mapping (UAM) and multi-scale similarity (MSS) across a comprehensive dataset of 14,012 CT scans. Furthermore, we developed MedLSAM by integrating MedLAM with the Segment Anything Model (SAM). This innovative framework requires extreme point annotations across three directions on several templates to enable MedLAM to locate the target anatomical structure in the image, with SAM performing the segmentation. It significantly reduces the amount of manual annotation required by SAM in 3D medical imaging scenarios. We conducted extensive experiments on two 3D datasets covering 38 distinct organs. Our findings are twofold: (1) MedLAM can directly localize anatomical structures using just a few template scans, achieving performance comparable to fully supervised models; (2) MedLSAM closely matches the performance of SAM and its specialized medical adaptations with manual prompts, while minimizing the need for extensive point annotations across the entire dataset. Moreover, MedLAM has the potential to be seamlessly integrated with future 3D SAM models, paving the way for enhanced segmentation performance. Our code is public at <span><span>https://github.com/openmedlab/MedLSAM</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"99 ","pages":"Article 103370"},"PeriodicalIF":10.7,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142503534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"HAGMN-UQ: Hyper association graph matching network with uncertainty quantification for coronary artery semantic labeling","authors":"Chen Zhao , Michele Esposito , Zhihui Xu , Weihua Zhou","doi":"10.1016/j.media.2024.103374","DOIUrl":"10.1016/j.media.2024.103374","url":null,"abstract":"<div><div>Coronary artery disease (CAD) is one of the leading causes of death worldwide. Accurate extraction of individual arterial branches from invasive coronary angiograms (ICA) is critical for CAD diagnosis and detection of stenosis. Generating semantic segmentation for coronary arteries through deep learning-based models presents challenges due to the morphological similarity among different types of coronary arteries, making it difficult to maintain high accuracy while keeping low computational complexity. To address this challenge, we propose an innovative approach using the hyper association graph-matching neural network with uncertainty quantification (HAGMN-UQ) for coronary artery semantic labeling on ICAs. The graph-matching procedure maps the arterial branches between two individual graphs, so that the unlabeled arterial segments are classified by the labeled segments, and the coronary artery semantic labeling is achieved. Leveraging hypergraphs not only extends representation capabilities beyond pairwise relationships, but also improves the robustness and accuracy of the graph matching by enabling the modeling of higher-order associations. In addition, employing the uncertainty quantification to determine the trustworthiness of graph matching reduces the required number of comparisons, so as to accelerate the inference speed. Consequently, our model achieved an accuracy of 0.9211 for coronary artery semantic labeling with a fast inference speed, leading to an effective and efficient prediction in real-time clinical decision-making scenarios.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"99 ","pages":"Article 103374"},"PeriodicalIF":10.7,"publicationDate":"2024-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142442814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Xianhua Zeng, Jianhua Gong, Weisheng Li, Zhuoya Yang
{"title":"Knowledge-driven multi-graph convolutional network for brain network analysis and potential biomarker discovery","authors":"Xianhua Zeng, Jianhua Gong, Weisheng Li, Zhuoya Yang","doi":"10.1016/j.media.2024.103368","DOIUrl":"10.1016/j.media.2024.103368","url":null,"abstract":"<div><div>In brain network analysis, individual-level data can provide biological features of individuals, while population-level data can provide demographic information of populations. However, existing methods mostly utilize either individual- or population-level features separately, inevitably neglecting the multi-level characteristics of brain disorders. To address this issue, we propose an end-to-end multi-graph neural network model called KMGCN. This model simultaneously leverages individual- and population-level features for brain network analysis. At the individual level, we construct multi-graph using both knowledge-driven and data-driven approaches. Knowledge-driven refers to constructing a knowledge graph based on prior knowledge, while data-driven involves learning a data graph from the data itself. At the population level, we construct multi-graph using both imaging and phenotypic data. Additionally, we devise a pooling method tailored for brain networks, capable of selecting brain regions that impact brain disorders. We evaluate the performance of our model on two large datasets, ADNI and ABIDE, and experimental results demonstrate that it achieves state-of-the-art performance, with 86.87% classification accuracy for ADNI and 86.40% for ABIDE, accompanied by around 10% improvements in all evaluation metrics compared to the state-of-the-art models. Additionally, the biomarkers identified by our model align well with recent neuroscience research, indicating the effectiveness of our model in brain network analysis and potential biomarker discovery. The code is available at <span><span>https://github.com/GN-gjh/KMGCN</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"99 ","pages":"Article 103368"},"PeriodicalIF":10.7,"publicationDate":"2024-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142442815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Samiksha Pachade , Prasanna Porwal , Manesh Kokare , Girish Deshmukh , Vivek Sahasrabuddhe , Zhengbo Luo , Feng Han , Zitang Sun , Li Qihan , Sei-ichiro Kamata , Edward Ho , Edward Wang , Asaanth Sivajohan , Saerom Youn , Kevin Lane , Jin Chun , Xinliang Wang , Yunchao Gu , Sixu Lu , Young-tack Oh , Fabrice Mériaudeau
{"title":"RFMiD: Retinal Image Analysis for multi-Disease Detection challenge","authors":"Samiksha Pachade , Prasanna Porwal , Manesh Kokare , Girish Deshmukh , Vivek Sahasrabuddhe , Zhengbo Luo , Feng Han , Zitang Sun , Li Qihan , Sei-ichiro Kamata , Edward Ho , Edward Wang , Asaanth Sivajohan , Saerom Youn , Kevin Lane , Jin Chun , Xinliang Wang , Yunchao Gu , Sixu Lu , Young-tack Oh , Fabrice Mériaudeau","doi":"10.1016/j.media.2024.103365","DOIUrl":"10.1016/j.media.2024.103365","url":null,"abstract":"<div><div>In the last decades, many publicly available large fundus image datasets have been collected for diabetic retinopathy, glaucoma, and age-related macular degeneration, and a few other frequent pathologies. These publicly available datasets were used to develop a computer-aided disease diagnosis system by training deep learning models to detect these frequent pathologies. One challenge limiting the adoption of a such system by the ophthalmologist is, computer-aided disease diagnosis system ignores sight-threatening rare pathologies such as central retinal artery occlusion or anterior ischemic optic neuropathy and others that ophthalmologists currently detect. Aiming to advance the state-of-the-art in automatic ocular disease classification of frequent diseases along with the rare pathologies, a grand challenge on “Retinal Image Analysis for multi-Disease Detection” was organized in conjunction with the IEEE International Symposium on Biomedical Imaging (ISBI - 2021). This paper, reports the challenge organization, dataset, top-performing participants solutions, evaluation measures, and results based on a new “Retinal Fundus Multi-disease Image Dataset” (RFMiD). There were two principal sub-challenges: disease screening (i.e. presence versus absence of pathology — a binary classification problem) and disease/pathology classification (a 28-class multi-label classification problem). It received a positive response from the scientific community with 74 submissions by individuals/teams that effectively entered in this challenge. The top-performing methodologies utilized a blend of data-preprocessing, data augmentation, pre-trained model, and model ensembling. This multi-disease (frequent and rare pathologies) detection will enable the development of generalizable models for screening the retina, unlike the previous efforts that focused on the detection of specific diseases.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"99 ","pages":"Article 103365"},"PeriodicalIF":10.7,"publicationDate":"2024-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142416675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yuliang Gu , Zhichao Sun , Tian Chen , Xin Xiao , Yepeng Liu , Yongchao Xu , Laurent Najman
{"title":"Dual structure-aware image filterings for semi-supervised medical image segmentation","authors":"Yuliang Gu , Zhichao Sun , Tian Chen , Xin Xiao , Yepeng Liu , Yongchao Xu , Laurent Najman","doi":"10.1016/j.media.2024.103364","DOIUrl":"10.1016/j.media.2024.103364","url":null,"abstract":"<div><div>Semi-supervised image segmentation has attracted great attention recently. The key is how to leverage unlabeled images in the training process. Most methods maintain consistent predictions of the unlabeled images under variations (<em>e.g.</em>, adding noise/perturbations, or creating alternative versions) in the image and/or model level. In most image-level variation, medical images often have prior structure information, which has not been well explored. In this paper, we propose novel dual structure-aware image filterings (DSAIF) as the image-level variations for semi-supervised medical image segmentation. Motivated by connected filtering that simplifies image via filtering in structure-aware tree-based image representation, we resort to the dual contrast invariant Max-tree and Min-tree representation. Specifically, we propose a novel connected filtering that removes topologically equivalent nodes (<em>i.e.</em> connected components) having no siblings in the Max/Min-tree. This results in two filtered images preserving topologically critical structure. Applying the proposed DSAIF to mutually supervised networks decreases the consensus of their erroneous predictions on unlabeled images. This helps to alleviate the confirmation bias issue of overfitting to noisy pseudo labels of unlabeled images, and thus effectively improves the segmentation performance. Extensive experimental results on three benchmark datasets demonstrate that the proposed method significantly/consistently outperforms some state-of-the-art methods. The source codes will be publicly available.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"99 ","pages":"Article 103364"},"PeriodicalIF":10.7,"publicationDate":"2024-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142442818","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Youjian Zhang , Li Li , Jie Wang , Xinquan Yang , Haotian Zhou , Jiahui He , Yaoqin Xie , Yuming Jiang , Wei Sun , Xinyuan Zhang , Guanqun Zhou , Zhicheng Zhang
{"title":"Texture-preserving diffusion model for CBCT-to-CT synthesis","authors":"Youjian Zhang , Li Li , Jie Wang , Xinquan Yang , Haotian Zhou , Jiahui He , Yaoqin Xie , Yuming Jiang , Wei Sun , Xinyuan Zhang , Guanqun Zhou , Zhicheng Zhang","doi":"10.1016/j.media.2024.103362","DOIUrl":"10.1016/j.media.2024.103362","url":null,"abstract":"<div><div>Cone beam computed tomography (CBCT) serves as a vital imaging modality in diverse clinical applications, but is constrained by inherent limitations such as reduced image quality and increased noise. In contrast, computed tomography (CT) offers superior resolution and tissue contrast. Bridging the gap between these modalities through CBCT-to-CT synthesis becomes imperative. Deep learning techniques have enhanced this synthesis, yet challenges with generative adversarial networks persist. Denoising Diffusion Probabilistic Models have emerged as a promising alternative in image synthesis. In this study, we propose a novel texture-preserving diffusion model for CBCT-to-CT synthesis that incorporates adaptive high-frequency optimization and a dual-mode feature fusion module. Our method aims to enhance high-frequency details, effectively fuse cross-modality features, and preserve fine image structures. Extensive validation demonstrates superior performance over existing methods, showcasing better generalization. The proposed model offers a transformative pathway to augment diagnostic accuracy and refine treatment planning across various clinical settings. This work represents a pivotal step toward non-invasive, safer, and high-quality CBCT-to-CT synthesis, advancing personalized diagnostic imaging practices.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"99 ","pages":"Article 103362"},"PeriodicalIF":10.7,"publicationDate":"2024-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142406521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}