Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention最新文献

筛选
英文 中文
Cortical Surface Parcellation using Spherical Convolutional Neural Networks. 使用球面卷积神经网络进行皮层表面分割。
Prasanna Parvathaneni, Shunxing Bao, Vishwesh Nath, Neil D Woodward, Daniel O Claassen, Carissa J Cascio, David H Zald, Yuankai Huo, Bennett A Landman, Ilwoo Lyu
{"title":"Cortical Surface Parcellation using Spherical Convolutional Neural Networks.","authors":"Prasanna Parvathaneni,&nbsp;Shunxing Bao,&nbsp;Vishwesh Nath,&nbsp;Neil D Woodward,&nbsp;Daniel O Claassen,&nbsp;Carissa J Cascio,&nbsp;David H Zald,&nbsp;Yuankai Huo,&nbsp;Bennett A Landman,&nbsp;Ilwoo Lyu","doi":"10.1007/978-3-030-32248-9_56","DOIUrl":"https://doi.org/10.1007/978-3-030-32248-9_56","url":null,"abstract":"<p><p>We present cortical surface parcellation using spherical deep convolutional neural networks. Traditional multi-atlas cortical surface parcellation requires inter-subject surface registration using geometric features with slow processing speed on a single subject (2-3 hours). Moreover, even optimal surface registration does not necessarily produce optimal cortical parcellation as parcel boundaries are not fully matched to the geometric features. In this context, a choice of training features is important for accurate cortical parcellation. To utilize the networks efficiently, we propose cortical parcellation-specific input data from an irregular and complicated structure of cortical surfaces. To this end, we align ground-truth cortical parcel boundaries and use their resulting deformation fields to generate new pairs of deformed geometric features and parcellation maps. To extend the capability of the networks, we then smoothly morph cortical geometric features and parcellation maps using the intermediate deformation fields. We validate our method on 427 adult brains for 49 labels. The experimental results show that our method outperforms traditional multi-atlas and naive spherical U-Net approaches, while achieving full cortical parcellation in less than a minute.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"11766 ","pages":"501-509"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6892466/pdf/nihms-1059107.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49687027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Active Appearance Model Induced Generative Adversarial Network for Controlled Data Augmentation. 用于受控数据增强的主动外观模型诱导的生成对抗性网络。
Jianfei Liu, Christine Shen, Tao Liu, Nancy Aguilera, Johnny Tam
{"title":"Active Appearance Model Induced Generative Adversarial Network for Controlled Data Augmentation.","authors":"Jianfei Liu,&nbsp;Christine Shen,&nbsp;Tao Liu,&nbsp;Nancy Aguilera,&nbsp;Johnny Tam","doi":"10.1007/978-3-030-32239-7_23","DOIUrl":"https://doi.org/10.1007/978-3-030-32239-7_23","url":null,"abstract":"<p><p>Data augmentation is an important strategy for enlarging training datasets in deep learning-based medical image analysis. This is because large, annotated medical datasets are not only difficult and costly to generate, but also quickly become obsolete due to rapid advances in imaging technology. Image-to-image conditional generative adversarial networks (C-GAN) provide a potential solution for data augmentation. However, annotations used as inputs to C-GAN are typically based only on shape information, which can result in undesirable intensity distributions in the resulting artificially-created images. In this paper, we introduce an active cell appearance model (ACAM) that can measure statistical distributions of shape and intensity and use this ACAM model to guide C-GAN to generate more realistic images, which we call A-GAN. A-GAN provides an effective means for conveying anisotropic intensity information to C-GAN. A-GAN incorporates a statistical model (ACAM) to determine how transformations are applied for data augmentation. Traditional approaches for data augmentation that are based on arbitrary transformations might lead to unrealistic shape variations in an augmented dataset that are not representative of real data. A-GAN is designed to ameliorate this. To validate the effectiveness of using A-GAN for data augmentation, we assessed its performance on cell analysis in adaptive optics retinal imaging, which is a rapidly-changing medical imaging modality. Compared to C-GAN, A-GAN achieved stability in fewer iterations. The cell detection and segmentation accuracy when assisted by A-GAN augmentation was higher than that achieved with C-GAN. These findings demonstrate the potential for A-GAN to substantially improve existing data augmentation methods in medical image analysis.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"11764 ","pages":"201-208"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6834374/pdf/nihms-1055537.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49687026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Pancreas Segmentation in MRI using Graph-Based Decision Fusion on Convolutional Neural Networks. 利用基于卷积神经网络的图式决策融合技术在核磁共振成像中进行胰腺分割
Jinzheng Cai, Le Lu, Zizhao Zhang, Fuyong Xing, Lin Yang, Qian Yin
{"title":"Pancreas Segmentation in MRI using Graph-Based Decision Fusion on Convolutional Neural Networks.","authors":"Jinzheng Cai, Le Lu, Zizhao Zhang, Fuyong Xing, Lin Yang, Qian Yin","doi":"10.1007/978-3-319-46723-8_51","DOIUrl":"https://doi.org/10.1007/978-3-319-46723-8_51","url":null,"abstract":"<p><p>Automated pancreas segmentation in medical images is a prerequisite for many clinical applications, such as diabetes inspection, pancreatic cancer diagnosis, and surgical planing. In this paper, we formulate pancreas segmentation in magnetic resonance imaging (MRI) scans as a graph based decision fusion process combined with deep convolutional neural networks (CNN). Our approach conducts pancreatic detection and boundary segmentation with two types of CNN models respectively: 1) the tissue detection step to differentiate pancreas and non-pancreas tissue with spatial intensity context; 2) the boundary detection step to allocate the semantic boundaries of pancreas. Both detection results of the two networks are fused together as the initialization of a conditional random field (CRF) framework to obtain the final segmentation output. Our approach achieves the mean dice similarity coefficient (DSC) 76.1% with the standard deviation of 8.7% in a dataset containing 78 abdominal MRI scans. The proposed algorithm achieves the best results compared with other state of the arts.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"9901 ","pages":"442-450"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5223591/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140195393","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust Cell Detection and Segmentation in Histopathological Images Using Sparse Reconstruction and Stacked Denoising Autoencoders. 使用稀疏重构和堆叠去噪自动编码器在组织病理学图像中进行稳健的细胞检测和分割
Hai Su, Fuyong Xing, Xiangfei Kong, Yuanpu Xie, Shaoting Zhang, Lin Yang
{"title":"Robust Cell Detection and Segmentation in Histopathological Images Using Sparse Reconstruction and Stacked Denoising Autoencoders.","authors":"Hai Su, Fuyong Xing, Xiangfei Kong, Yuanpu Xie, Shaoting Zhang, Lin Yang","doi":"10.1007/978-3-319-24574-4_46","DOIUrl":"https://doi.org/10.1007/978-3-319-24574-4_46","url":null,"abstract":"<p><p>Computer-aided diagnosis (CAD) is a promising tool for accurate and consistent diagnosis and prognosis. Cell detection and segmentation are essential steps for CAD. These tasks are challenging due to variations in cell shapes, touching cells, and cluttered background. In this paper, we present a cell detection and segmentation algorithm using the sparse reconstruction with trivial templates and a stacked denoising autoencoder (sDAE). The sparse reconstruction handles the shape variations by representing a testing patch as a linear combination of shapes in the learned dictionary. Trivial templates are used to model the touching parts. The sDAE, trained with the original data and their structured labels, is used for cell segmentation. To the best of our knowledge, this is the first study to apply sparse reconstruction and sDAE with structured labels for cell detection and segmentation. The proposed method is extensively tested on two data sets containing more than 3000 cells obtained from brain tumor and lung cancer images. Our algorithm achieves the best performance compared with other state of the arts.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"9351 ","pages":"383-390"},"PeriodicalIF":0.0,"publicationDate":"2015-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5081214/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140290109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated Model-Based Segmentation of the Left and Right Ventricles in Tagged Cardiac MRI. 标记心脏MRI中左心室和右心室的自动模型分割。
Albert Montillo, Dimitris Metaxas, Leon Axel
{"title":"Automated Model-Based Segmentation of the Left and Right Ventricles in Tagged Cardiac MRI.","authors":"Albert Montillo,&nbsp;Dimitris Metaxas,&nbsp;Leon Axel","doi":"10.1007/978-3-540-39899-8_63","DOIUrl":"https://doi.org/10.1007/978-3-540-39899-8_63","url":null,"abstract":"<p><p>We describe an automated, model-based method to segment the left and right ventricles in 4D tagged MR. We fit 3D epicardial and endocardial surface models to ventricle features we extract from the image data. Excellent segmentation is achieved using novel methods that (1) initialize the models and (2) that compute 3D model forces from 2D tagged MR images. The 3D forces guide the models to patient-specific anatomy while the fit is regularized via internal deformation strain energy of a thin plate. Deformation continues until the forces equilibrate or vanish. Validation of the segmentations is performed quantitatively and qualitatively on normal and diseased subjects.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"2878 ","pages":"507-515"},"PeriodicalIF":0.0,"publicationDate":"2003-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1007/978-3-540-39899-8_63","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41224576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Automated Segmentation of the Left and Right Ventricles in 4D Cardiac SPAMM Images. 4D心脏SPAMM图像中左心室和右心室的自动分割。
Albert Montillo, Dimitris Metaxas, Leon Axel
{"title":"Automated Segmentation of the Left and Right Ventricles in 4D Cardiac SPAMM Images.","authors":"Albert Montillo,&nbsp;Dimitris Metaxas,&nbsp;Leon Axel","doi":"10.1007/3-540-45786-0_77","DOIUrl":"https://doi.org/10.1007/3-540-45786-0_77","url":null,"abstract":"<p><p>In this paper we describe a completely automated volume-based method for the segmentation of the left and right ventricles in 4D tagged MR (SPAMM) images for quantitative cardiac analysis. We correct the background intensity variation in each volume caused by surface coils using a new scale-based fuzzy connectedness procedure. We apply 3D grayscale opening to the corrected data to create volumes containing only the blood filled regions. We threshold the volumes by minimizing region variance or by an adaptive statistical thresholding method. We isolate the ventricular blood filled regions using a novel approach based on spatial and temporal shape similarity. We use these regions to define the endocardium contours and use them to initialize an active contour that locates the epicardium through the gradient vector flow of an edgemap of a grayscale-closed image. Both quantitative and qualitative results on normal and diseased patients are presented.</p>","PeriodicalId":94280,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"2488 ","pages":"620-633"},"PeriodicalIF":0.0,"publicationDate":"2002-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1007/3-540-45786-0_77","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49687025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 59
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信