Haleh Akrami, Anand A Joshi, Sergül Aydöre, Richard M Leahy
{"title":"Deep Quantile Regression for Uncertainty Estimation in Unsupervised and Supervised Lesion Detection.","authors":"Haleh Akrami, Anand A Joshi, Sergül Aydöre, Richard M Leahy","doi":"","DOIUrl":"","url":null,"abstract":"<p><p>Despite impressive state-of-the-art performance on a wide variety of machine learning tasks in multiple applications, deep learning methods can produce over-confident predictions, particularly with limited training data. Therefore, quantifying uncertainty is particularly important in critical applications such as lesion detection and clinical diagnosis, where a realistic assessment of uncertainty is essential in determining surgical margins, disease status and appropriate treatment. In this work, we propose a novel approach that uses quantile regression for quantifying aleatoric uncertainty in both supervised and unsupervised lesion detection problems. The resulting confidence intervals can be used for lesion detection and segmentation. In the unsupervised setting, we combine quantile regression with the Variational AutoEncoder (VAE). The VAE is trained on lesion-free data, so when presented with an image with a lesion, it tends to reconstruct a lesion-free version of the image. To detect the lesion, we then compare the input (lesion) and output (lesion-free) images. Here we address the problem of quantifying uncertainty in the images that are reconstructed by the VAE as the basis for principled outlier or lesion detection. The VAE models the output as a conditionally independent Gaussian characterized by its mean and variance. Unfortunately, joint optimization of both mean and variance in the VAE leads to the well-known problem of shrinkage or underestimation of variance. Here we describe an alternative Quantile-Regression VAE (QR-VAE) that avoids this variance shrinkage problem by directly estimating conditional quantiles for the input image. Using the estimated quantiles, we compute the conditional mean and variance for the input image from which we then detect outliers by thresholding at a false-discovery-rate corrected p-value. In the supervised setting, we develop binary quantile regression (BQR) for the supervised lesion segmentation task. We show how BQR can be used to capture uncertainty in lesion boundaries in a manner that characterizes expert disagreement.</p>","PeriodicalId":75083,"journal":{"name":"The journal of machine learning for biomedical imaging","volume":"1 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9881592/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10646814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Raghav Mehta, Angelos Filos, Ujjwal Baid, C. Sako, Richard McKinley, M. Rebsamen, K. Datwyler, Raphael Meier, P. Radojewski, G. Murugesan, S. Nalawade, Chandan Ganesh, B. Wagner, F. Yu, B. Fei, A. Madhuranthakam, J. Maldjian, L. Daza, Catalina G'omez, P. Arbel'aez, Chengliang Dai, Shuo Wang, Hadrien Raynaud, Yuanhan Mo, E. Angelini, Yike Guo, Wenjia Bai, Subhashis Banerjee, L. Pei, A. Murat, Sarahi Rosas-Gonz'alez, Illyess Zemmoura, C. Tauber, Minh H. Vu, T. Nyholm, T. Lofstedt, Laura Mora Ballestar, Verónica Vilaplana, Hugh McHugh, G. M. Talou, Alan Wang, J. Patel, Ken Chang, K. Hoebel, M. Gidwani, N. Arun, Sharut Gupta, M. Aggarwal, Praveer Singh, E. Gerstner, Jayashree Kalpathy-Cramer, Nicolas Boutry, Alexis Huard, L. Vidyaratne, Md Monibor Rahman, K. Iftekharuddin, J. Chazalon, É. Puybareau, G. Tochon, Jun Ma, M. Cabezas, X. Lladó, A. Oliver, Liliana Valencia, S. Valverde, Mehdi Amian, M. Soltaninejad, A. Myronenko, Ali Hatamizadeh, Xuejing Feng, Q. Dou, N. Tustison, Craig Meyer, Nisarg A. Shah, S. Ta
{"title":"QU-BraTS: MICCAI BraTS 2020 Challenge on Quantifying Uncertainty in Brain Tumor Segmentation - Analysis of Ranking Metrics and Benchmarking Results","authors":"Raghav Mehta, Angelos Filos, Ujjwal Baid, C. Sako, Richard McKinley, M. Rebsamen, K. Datwyler, Raphael Meier, P. Radojewski, G. Murugesan, S. Nalawade, Chandan Ganesh, B. Wagner, F. Yu, B. Fei, A. Madhuranthakam, J. Maldjian, L. Daza, Catalina G'omez, P. Arbel'aez, Chengliang Dai, Shuo Wang, Hadrien Raynaud, Yuanhan Mo, E. Angelini, Yike Guo, Wenjia Bai, Subhashis Banerjee, L. Pei, A. Murat, Sarahi Rosas-Gonz'alez, Illyess Zemmoura, C. Tauber, Minh H. Vu, T. Nyholm, T. Lofstedt, Laura Mora Ballestar, Verónica Vilaplana, Hugh McHugh, G. M. Talou, Alan Wang, J. Patel, Ken Chang, K. Hoebel, M. Gidwani, N. Arun, Sharut Gupta, M. Aggarwal, Praveer Singh, E. Gerstner, Jayashree Kalpathy-Cramer, Nicolas Boutry, Alexis Huard, L. Vidyaratne, Md Monibor Rahman, K. Iftekharuddin, J. Chazalon, É. Puybareau, G. Tochon, Jun Ma, M. Cabezas, X. Lladó, A. Oliver, Liliana Valencia, S. Valverde, Mehdi Amian, M. Soltaninejad, A. Myronenko, Ali Hatamizadeh, Xuejing Feng, Q. Dou, N. Tustison, Craig Meyer, Nisarg A. Shah, S. Ta","doi":"10.59275/j.melba.2022-354b","DOIUrl":"https://doi.org/10.59275/j.melba.2022-354b","url":null,"abstract":"Deep learning (DL) models have provided state-of-the-art performance in various medical imaging benchmarking challenges, including the Brain Tumor Segmentation (BraTS) challenges. However, the task of focal pathology multi-compartment segmentation (e.g., tumor and lesion sub-regions) is particularly challenging, and potential errors hinder translating DL models into clinical workflows. Quantifying the reliability of DL model predictions in the form of uncertainties could enable clinical review of the most uncertain regions, thereby building trust and paving the way toward clinical translation. Several uncertainty estimation methods have recently been introduced for DL medical image segmentation tasks. Developing scores to evaluate and compare the performance of uncertainty measures will assist the end-user in making more informed decisions. In this study, we explore and evaluate a score developed during the BraTS 2019 and BraTS 2020 task on uncertainty quantification (QU-BraTS) and designed to assess and rank uncertainty estimates for brain tumor multi-compartment segmentation. This score (1) rewards uncertainty estimates that produce high confidence in correct assertions and those that assign low confidence levels at incorrect assertions, and (2) penalizes uncertainty measures that lead to a higher percentage of under-confident correct assertions. We further benchmark the segmentation uncertainties generated by 14 independent participating teams of QU-BraTS 2020, all of which also participated in the main BraTS segmentation task. Overall, our findings confirm the importance and complementary value that uncertainty estimates provide to segmentation algorithms, highlighting the need for uncertainty quantification in medical image analyses. Finally, in favor of transparency and reproducibility, our evaluation code is made publicly available at https://github.com/RagMeh11/QU-BraTS.","PeriodicalId":75083,"journal":{"name":"The journal of machine learning for biomedical imaging","volume":"86 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85871515","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
H. Akrami, Anand A. Joshi, Sergül Aydöre, R. Leahy
{"title":"Deep Quantile Regression for Uncertainty Estimation in Unsupervised and Supervised Lesion Detection","authors":"H. Akrami, Anand A. Joshi, Sergül Aydöre, R. Leahy","doi":"10.59275/j.melba.2022-6751","DOIUrl":"https://doi.org/10.59275/j.melba.2022-6751","url":null,"abstract":"Despite impressive state-of-the-art performance on a wide variety of machine learning tasks in multiple applications, deep learning methods can produce over-confident predictions, particularly with limited training data. Therefore, quantifying uncertainty is particularly important in critical applications such as lesion detection and clinical diagnosis, where a realistic assessment of uncertainty is essential in determining surgical margins, disease status and appropriate treatment. In this work, we propose a novel approach that uses quantile regression for quantifying aleatoric uncertainty in both supervised and unsupervised lesion detection problems. The resulting confidence intervals can be used for lesion detection and segmentation. In the unsupervised setting, we combine quantile regression with the Variational AutoEncoder (VAE). The VAE is trained on lesion-free data, so when presented with an image with a lesion, it tends to reconstruct a lesion-free version of the image. To detect the lesion, we then compare the input (lesion) and output (lesion-free) images. Here we address the problem of quantifying uncertainty in the images that are reconstructed by the VAE as the basis for principled outlier or lesion detection. The VAE models the output as a conditionally independent Gaussian characterized by its mean and variance. Unfortunately, joint optimization of both mean and variance in the VAE leads to the well-known problem of shrinkage or underestimation of variance. Here we describe an alternative Quantile-Regression VAE (QR-VAE) that avoids this variance shrinkage problem by directly estimating conditional quantiles for the input image. Using the estimated quantiles, we compute the conditional mean and variance for the input image from which we then detect outliers by thresholding at a false-discovery-rate corrected p-value. In the supervised setting, we develop binary quantile regression (BQR) for the supervised lesion segmentation task. We show how BQR can be used to capture uncertainty in lesion boundaries in a manner that characterizes expert disagreement.","PeriodicalId":75083,"journal":{"name":"The journal of machine learning for biomedical imaging","volume":"37 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89982900","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A review and experimental evaluation of deep learning methods for MRI reconstruction","authors":"Arghya Pal, Y. Rathi","doi":"10.59275/j.melba.2022-3g12","DOIUrl":"https://doi.org/10.59275/j.melba.2022-3g12","url":null,"abstract":"Following the success of deep learning in a wide range of applications, neural network-based machine-learning techniques have received significant interest for accelerating magnetic resonance imaging (MRI) acquisition and reconstruction strategies. A number of ideas inspired by deep learning techniques for computer vision and image processing have been successfully applied to nonlinear image reconstruction in the spirit of compressed sensing for accelerated MRI. Given the rapidly growing nature of the field, it is imperative to consolidate and summarize the large number of deep learning methods that have been reported in the literature, to obtain a better understanding of the field in general. This article provides an overview of the recent developments in neural-network based approaches that have been proposed specifically for improving parallel imaging. A general background and introduction to parallel MRI is also given from a classical view of k-space based reconstruction methods. Image domain based techniques that introduce improved regularizers are covered along with k-space based methods which focus on better interpolation strategies using neural networks. While the field is rapidly evolving with plenty of papers published each year, in this review, we attempt to cover broad categories of methods that have shown good performance on publicly available data sets. Limitations and open problems are also discussed and recent efforts for producing open data sets and benchmarks for the community are examined.","PeriodicalId":75083,"journal":{"name":"The journal of machine learning for biomedical imaging","volume":"8 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77744599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Raghavendra Selvan, E. Dam, Soren Alexander Flensborg, Jens Petersen
{"title":"Patch-based Medical Image Segmentation using Matrix Product State Tensor Networks","authors":"Raghavendra Selvan, E. Dam, Soren Alexander Flensborg, Jens Petersen","doi":"10.59275/j.melba.2022-d1f5","DOIUrl":"https://doi.org/10.59275/j.melba.2022-d1f5","url":null,"abstract":"Tensor networks are efficient factorisations of high dimensional tensors into network of lower order tensors. They have been most commonly used to model entanglement in quantum many-body systems and more recently are witnessing increased applications in supervised machine learning. In this work, we formulate image segmentation in a supervised setting with tensor networks. The key idea is to first lift the pixels in image patches to exponentially high dimensional feature spaces and using a linear decision hyper-plane to classify the input pixels into foreground and background classes. The high dimensional linear model itself is approximated using the matrix product state (MPS) tensor network. The MPS is weight-shared between the non-overlapping image patches resulting in our strided tensor network model. The performance of the proposed model is evaluated on three three 2D- and one 3D- biomedical imaging datasets. The performance of the proposed tensor network segmentation model is compared with relevant baseline methods. In the 2D experiments, the tensor network model yeilds competitive performance compared to the baseline methods while being more resource efficient.","PeriodicalId":75083,"journal":{"name":"The journal of machine learning for biomedical imaging","volume":"18 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75450039","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
K. Arnavaz, Oswin Krause, Kilian Zepf, J. A. Bærentzen, Jelena M. Krivokapic, Silja Heilmann, P. Nyeng, Aasa Feragen
{"title":"Quantifying Topology In Pancreatic Tubular Networks From Live Imaging 3D Microscopy","authors":"K. Arnavaz, Oswin Krause, Kilian Zepf, J. A. Bærentzen, Jelena M. Krivokapic, Silja Heilmann, P. Nyeng, Aasa Feragen","doi":"10.59275/j.melba.2022-4bf2","DOIUrl":"https://doi.org/10.59275/j.melba.2022-4bf2","url":null,"abstract":"Motivated by the challenging segmentation task of pancreatic tubular networks, this paper tackles two commonly encountered problems in biomedical imaging: Topological consistency of the segmentation, and expensive or difficult annotation. Our contributions are the following: a) We propose a topological score which measures both topological and geometric consistency between the predicted and ground truth segmentations, applied to model selection and validation. b) We provide a full deep-learning methodology for this difficult noisy task on time-series image data. In our method, we first use a semisupervised U-net architecture, applicable to generic segmentation tasks, which jointly trains an autoencoder and a segmentation network. We then use tracking of loops over time to further improve the predicted topology. This semi-supervised approach allows us to utilize unannotated data to learn feature representations that generalize to test data with high variability, in spite of our annotated training data having very limited variation. Our contributions are validated on a challenging segmentation task, locating tubular structures in the fetal pancreas from noisy live imaging confocal microscopy. We show that our semi-supervised model outperforms not only fully supervised and pre-trained models but also an approach which takes topological consistency into account during training. Further, our approach achieves a mean loop score of 0.808 for detecting loops in the fetal pancreas, compared to a U-net trained with clDice with mean loop score 0.762.","PeriodicalId":75083,"journal":{"name":"The journal of machine learning for biomedical imaging","volume":"18 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78869543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Semi-Supervised Federated Peer Learning for Skin Lesion Classification","authors":"T. Bdair, N. Navab, Shadi Albarqouni","doi":"10.59275/j.melba.2022-8g82","DOIUrl":"https://doi.org/10.59275/j.melba.2022-8g82","url":null,"abstract":"Globally, Skin carcinoma is among the most lethal diseases. Millions of people are diagnosed with this cancer every year. Sill, early detection can decrease the medication cost and mortality rate substantially. The recent improvement in automated cancer classification using deep learning methods has reached a human-level performance requiring a large amount of annotated data assembled in one location, yet, finding such conditions usually is not feasible. Recently, federated learning (FL) has been proposed to train decentralized models in a privacy-preserved fashion depending on labeled data at the client-side, which is usually not available and costly. To address this, we propose FedPerl, a semi-supervised federated learning method. Our method is inspired by peer learning from educational psychology and ensemble averaging from committee machines. FedPerl builds communities based on clients' similarities. Then it encourages communities' members to learn from each other to generate more accurate pseudo labels for the unlabeled data. We also proposed the peer anonymization (PA) technique to anonymize clients. As a core component of our method, PA is orthogonal to other methods without additional complexity, and reduces the communication cost while enhances performance. Finally, we propose a dynamic peer learning policy that controls the learning stream to avoid any degradation in the performance, especially for the individual clients. Our experimental setup consists of 71,000 skin lesion images collected from 5 publicly available datasets. We test our method in four different scenarios in SSFL. With few annotated data, FedPerl is on par with a state-of-the-art method in skin lesion classification in the standard setup while outperforming SSFLs and the baselines by 1.8% and 15.8%, respectively. Also, it generalizes better to an unseen client while being less sensitive to noisy ones.","PeriodicalId":75083,"journal":{"name":"The journal of machine learning for biomedical imaging","volume":"7 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74951977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Francesco Calivá, Kaiyang Cheng, Rutwik Shah, V. Pedoia
{"title":"Adversarial Robust Training of Deep Learning MRI Reconstruction Models","authors":"Francesco Calivá, Kaiyang Cheng, Rutwik Shah, V. Pedoia","doi":"10.59275/j.melba.2021-df47","DOIUrl":"https://doi.org/10.59275/j.melba.2021-df47","url":null,"abstract":"Deep Learning (DL) has shown potential in accelerating Magnetic Resonance Image acquisition and reconstruction. Nevertheless, there is a dearth of tailored methods to guarantee that the reconstruction of small features is achieved with high fidelity. In this work, we employ adversarial attacks to generate small synthetic perturbations, which are difficult to reconstruct for a trained DL reconstruction network. Then, we use robust training to increase the network’s sensitivity to these small features and encourage their reconstruction. Next, we investigate the generalization of said approach to real world features. For this, a musculoskeletal radiologist annotated a set of cartilage and meniscal lesions from the knee Fast-MRI dataset, and a classification network was devised to assess the reconstruction of the features. Experimental results show that by introducing robust training to a reconstruction network, the rate of false negative features (4.8%) in image reconstruction can be reduced. These results are encouraging, and highlight the necessity for attention to this problem by the image reconstruction community, as a milestone for the introduction of DL reconstruction in clinical practice. To support further research, we make our annotations and code publicly available at https://github.com/fcaliva/fastMRI_BB_abnormalities_annotation.","PeriodicalId":75083,"journal":{"name":"The journal of machine learning for biomedical imaging","volume":"33 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81011312","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Nested Grassmannians for Dimensionality Reduction with Applications","authors":"Chun-Hao Yang, B. Vemuri","doi":"10.59275/j.melba.2022-234f","DOIUrl":"https://doi.org/10.59275/j.melba.2022-234f","url":null,"abstract":"In the recent past, nested structures in Riemannian manifolds has been studied in the context of dimensionality reduction as an alternative to the popular principal geodesic analysis (PGA) technique, for example, the principal nested spheres. In this paper, we propose a novel framework for constructing a nested sequence of homogeneous Riemannian manifolds. Common examples of homogeneous Riemannian manifolds include the n-sphere, the Stiefel manifold, the Grassmann manifold and many others. In particular, we focus on applying the proposed framework to the Grassmann manifold, giving rise to the nested Grassmannians (NG). An important application in which Grassmann manifolds are encountered is planar shape analysis. Specifically, each planar (2D) shape can be represented as a point in the complex projective space which is a complex Grassmann manifold. Some salient features of our framework are: (i) it explicitly exploits the geometry of the homogeneous Riemannian manifolds and (ii) the nested lower-dimensional submanifolds need not be geodesic. With the proposed NG structure, we develop algorithms for the supervised and unsupervised dimensionality reduction problems respectively. The proposed algorithms are compared with PGA via simulation studies and real data experiments and are shown to achieve a higher ratio of expressed variance compared to PGA.","PeriodicalId":75083,"journal":{"name":"The journal of machine learning for biomedical imaging","volume":"9 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74889425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jinwei Zhang, Hang Zhang, M. Sabuncu, P. Spincemaille, Thanh D. Nguyen, Yi Wang
{"title":"Probabilistic dipole inversion for adaptive quantitative susceptibility mapping","authors":"Jinwei Zhang, Hang Zhang, M. Sabuncu, P. Spincemaille, Thanh D. Nguyen, Yi Wang","doi":"10.59275/j.melba.2021-bbf2","DOIUrl":"https://doi.org/10.59275/j.melba.2021-bbf2","url":null,"abstract":"A learning-based posterior distribution estimation method, Probabilistic Dipole Inversion (PDI), is proposed to solve the quantitative susceptibility mapping (QSM) inverse problem in MRI with uncertainty estimation. In PDI, a deep convolutional neural network (CNN) is used to represent the multivariate Gaussian distribution as the approximate posterior distribution of susceptibility given the input measured field. Such CNN is first trained on healthy subjects via posterior density estimation, where the training dataset contains samples from the true posterior distribution. Domain adaptations are then deployed on patient datasets with new pathologies not included in pre-training, where PDI updates the pre-trained CNN’s weights in an unsupervised fashion by minimizing the Kullback-Leibler divergence between the approximate posterior distribution represented by CNN and the true posterior distribution from the likelihood distribution of a known physical model and pre-defined prior distribution. Based on our experiments, PDI provides additional uncertainty estimation compared to the conventional MAP approach, while addressing the potential issue of the pre-trained CNN when test data deviates from training. Our code is available at https://github.com/Jinwei1209/Bayesian_QSM.","PeriodicalId":75083,"journal":{"name":"The journal of machine learning for biomedical imaging","volume":"745 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2020-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76841908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}