Journal of Medical Imaging最新文献

筛选
英文 中文
LiteMIL: a computationally efficient cross-attention multiple instance learning for cancer subtyping on whole-slide images. LiteMIL:一种计算效率高的跨注意多实例学习,用于整张幻灯片图像的癌症亚型分型。
IF 1.7
Journal of Medical Imaging Pub Date : 2026-01-01 Epub Date: 2026-01-06 DOI: 10.1117/1.JMI.13.1.017501
Haitham Kussaibi
{"title":"LiteMIL: a computationally efficient cross-attention multiple instance learning for cancer subtyping on whole-slide images.","authors":"Haitham Kussaibi","doi":"10.1117/1.JMI.13.1.017501","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.017501","url":null,"abstract":"<p><strong>Purpose: </strong>Accurate cancer subtyping is essential for precision medicine but challenged by the computational demands of gigapixel whole-slide images (WSIs). Although transformer-based multiple instance learning (MIL) methods achieve strong performance, their quadratic complexity limits clinical deployment. We introduce LiteMIL, a computationally efficient cross-attention MIL, optimized for WSIs classification.</p><p><strong>Approach: </strong>LiteMIL employs a single learnable query with multi-head cross-attention for bag-level aggregation from extracted features. We evaluated LiteMIL against five baselines (mean/max pooling, ABMIL, MAD-MIL, and TransMIL) on four TCGA datasets (breast: <math><mrow><mi>n</mi> <mo>=</mo> <mn>875</mn></mrow> </math> , kidney: <math><mrow><mi>n</mi> <mo>=</mo> <mn>906</mn></mrow> </math> , lung: <math><mrow><mi>n</mi> <mo>=</mo> <mn>958</mn></mrow> </math> , and TUPAC16: <math><mrow><mi>n</mi> <mo>=</mo> <mn>821</mn></mrow> </math> ) using nested cross-validation with patient-level splitting. Systematic ablation studies evaluated multi-query variants, attention heads, dropout rates, and architectural components.</p><p><strong>Results: </strong>LiteMIL achieved competitive accuracy (average 83.5%), matching TransMIL, while offering substantial efficiency gains: <math><mrow><mn>4.8</mn> <mo>×</mo></mrow> </math> fewer parameters (560K versus 2.67M), <math><mrow><mn>2.9</mn> <mo>×</mo></mrow> </math> faster inference (1.6s versus 4.6s per fold), and <math><mrow><mn>6.7</mn> <mo>×</mo></mrow> </math> lower Graphics Processing Unit (GPU) memory usage (1.15 GB versus 7.77 GB). LiteMIL excelled on lung (86.3% versus 85.0%), TUPAC16 (72% versus 71.4%), and matched kidney performance (89.9% versus 89.7%). Ablation studies revealed task-dependent multi-query performance benefits: <math><mrow><mi>Q</mi> <mo>=</mo> <mn>4</mn></mrow> </math> versus <math><mrow><mi>Q</mi> <mo>=</mo> <mn>1</mn></mrow> </math> improved morphologically heterogeneous tasks (breast/lung <math><mrow><mo>+</mo> <mn>1.3</mn> <mo>%</mo></mrow> </math> each, <math><mrow><mi>p</mi> <mo><</mo> <mn>0.05</mn></mrow> </math> ) but degraded on grading tasks (TUPAC16: <math><mrow><mo>-</mo> <mn>1.6</mn> <mo>%</mo></mrow> </math> ), validating single-query optimality for focused attention scenarios.</p><p><strong>Conclusions: </strong>LiteMIL provides a resource-efficient solution for WSI classification. The cross-attention architecture matches complex transformer performance while enabling deployment on consumer GPUs. Task-dependent design insights, single query for sparse discriminating features, multi-query for heterogeneous patterns, guide practical implementation. The architecture's efficiency, combined with compact features, makes LiteMIL suitable for clinical integration in settings with limited computational infrastructure.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"017501"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12772520/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145918838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Asynchronous federated learning for web-based OCT image analysis. 基于web的OCT图像分析异步联合学习。
IF 1.7
Journal of Medical Imaging Pub Date : 2026-01-01 Epub Date: 2026-01-05 DOI: 10.1117/1.JMI.13.1.014501
Hasan Md Tusfiqur Alam, Tim Maurer, Abdulrahman Mohamed Selim, Matthias Eiletz, Michael Barz, Daniel Sonntag
{"title":"Asynchronous federated learning for web-based OCT image analysis.","authors":"Hasan Md Tusfiqur Alam, Tim Maurer, Abdulrahman Mohamed Selim, Matthias Eiletz, Michael Barz, Daniel Sonntag","doi":"10.1117/1.JMI.13.1.014501","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014501","url":null,"abstract":"<p><strong>Purpose: </strong>Centralized machine learning often struggles with limited data access and expert involvement. We investigate decentralized approaches that preserve data privacy while enabling collaborative model training for medical imaging tasks.</p><p><strong>Approach: </strong>We explore asynchronous federated learning (FL) using the FL with buffered asynchronous aggregation (FedBuff) algorithm for classifying optical coherence tomography (OCT) retina images. Unlike synchronous algorithms such as FedAvg, which require all clients to participate simultaneously, FedBuff supports independent client updates. We compare its performance to both centralized models and FedAvg. In addition, we develop a browser-based proof-of-concept system using modern web technologies to assess the feasibility and limitations of interactive, collaborative learning in real-world settings.</p><p><strong>Results: </strong>FedBuff performs well in binary OCT classification tasks but shows reduced accuracy in more complex, multiclass scenarios. FedAvg achieves results comparable to centralized training, consistent with previous findings. Although FedBuff underperforms compared with FedAvg and centralized models, it still delivers acceptable accuracy in less complex settings. The browser-based prototype demonstrates the potential for accessible, user-driven FL systems but also highlights technical limitations in current web standards, especially regarding local computation and communication efficiency.</p><p><strong>Conclusion: </strong>Asynchronous FL via FedBuff offers a promising, privacy-preserving approach for medical image classification, particularly when synchronous participation is impractical. However, its scalability to complex classification tasks remains limited. Web-based implementations have the potential to broaden access to collaborative AI tools, but limitations of the current technologies need to be further investigated.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014501"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12767620/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145913135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Characterizing the effects of noncontrast head CT reconstruction kernel and slice thickness parameters on the performance of an automated AI algorithm in the evaluation of ischemic stroke. 表征非对比头部CT重建核和层厚参数对一种评估缺血性卒中的自动AI算法性能的影响。
IF 1.7
Journal of Medical Imaging Pub Date : 2026-01-01 Epub Date: 2026-01-09 DOI: 10.1117/1.JMI.13.1.014503
Spencer H Welland, Grace Hyun J Kim, Anil Yadav, Kambiz Nael, John M Hoffman, Matthew S Brown, Michael F McNitt-Gray, William Hsu
{"title":"Characterizing the effects of noncontrast head CT reconstruction kernel and slice thickness parameters on the performance of an automated AI algorithm in the evaluation of ischemic stroke.","authors":"Spencer H Welland, Grace Hyun J Kim, Anil Yadav, Kambiz Nael, John M Hoffman, Matthew S Brown, Michael F McNitt-Gray, William Hsu","doi":"10.1117/1.JMI.13.1.014503","DOIUrl":"10.1117/1.JMI.13.1.014503","url":null,"abstract":"<p><strong>Purpose: </strong>There are multiple commercially available, Food and Drug Administration (FDA)-cleared, artificial intelligence (AI)-based tools automating stroke evaluation in noncontrast computed tomography (NCCT). This study assessed the impact of variations in reconstruction kernel and slice thickness on two outputs of such a system: hypodense volume and Alberta Stroke Program Early CT Score (ASPECTS).</p><p><strong>Approach: </strong>The NCCT series image data of 67 patients imaged with a CT stroke protocol were reconstructed with four kernels (H10s-smooth, H40s-medium, H60s-sharp, and H70h-very sharp) and three slice thicknesses (1.5, 3.0, and 5.0 mm) to create 1 reference condition (H40s/5.0 mm) and 11 nonreference conditions. The 12 reconstructions per patient were processed with a commercially available FDA-cleared software package that yields total hypodense volume (mL) and ASPECTS. A mixed-effect model was used to test the difference in hypodense volume, and an ordered logistic model was used to test the difference in e-ASPECTS.</p><p><strong>Results: </strong>Hypodense volume differences from the reference condition ranged from <math><mrow><mo>-</mo> <mn>14.6</mn></mrow> </math> to 1.1 mL and were significant for all nonreference kernels (H10s <math><mrow><mi>p</mi> <mo>=</mo> <mn>0.025</mn></mrow> </math> , H60s <math><mrow><mi>p</mi> <mo><</mo> <mn>0.001</mn></mrow> </math> , and H70h <math><mrow><mi>p</mi> <mo><</mo> <mn>0.001</mn></mrow> </math> ) and for thinner slices (1.5 mm <math><mrow><mi>p</mi> <mo><</mo> <mn>0.001</mn></mrow> </math> and 3.0 mm <math><mrow><mi>p</mi> <mo>=</mo> <mn>0.002</mn></mrow> </math> ). e-ASPECTS was invariant to the nonreference kernels and slice thicknesses, with a mean difference ranging from <math><mrow><mo>-</mo> <mn>0.1</mn></mrow> </math> to 0.5. No significant differences were found for any kernel or slice thickness (all <math><mrow><mi>p</mi> <mo>></mo> <mn>0.05</mn></mrow> </math> ).</p><p><strong>Conclusions: </strong>Automated hypodense volume measured with a commercially available, FDA-cleared software package is substantially impacted by reconstruction kernel and slice thickness. Conversely, automated ASPECTS is invariant to these reconstruction parameters.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014503"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12782953/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145953582","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Utility of the virtual imaging trials methodology for objective characterization of AI systems and training data. 利用虚拟成像试验方法客观表征人工智能系统和训练数据。
IF 1.7
Journal of Medical Imaging Pub Date : 2026-01-01 Epub Date: 2026-02-15 DOI: 10.1117/1.JMI.13.1.014506
Fakrul Islam Tushar, Lavsen Dahal, Saman Sotoudeh-Paima, Ehsan Abadi, William P Segars, Joseph Y Lo, Ehsan Samei
{"title":"Utility of the virtual imaging trials methodology for objective characterization of AI systems and training data.","authors":"Fakrul Islam Tushar, Lavsen Dahal, Saman Sotoudeh-Paima, Ehsan Abadi, William P Segars, Joseph Y Lo, Ehsan Samei","doi":"10.1117/1.JMI.13.1.014506","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014506","url":null,"abstract":"<p><strong>Purpose: </strong>The credibility of artificial intelligence (AI) models for medical imaging continues to be a challenge, affected by the diversity of models, the data used to train the models, and the applicability of their combination to produce reproducible results for new data. We aimed to explore whether emerging virtual imaging trial (VIT) methodologies can provide an objective resource to approach this challenge.</p><p><strong>Approach: </strong>We conducted this study for the case example of COVID-19 diagnosis using clinical and virtual computed tomography (CT) and chest radiography (CXR) processed with convolutional neural networks. Multiple AI models were developed and tested using 3D ResNet-like and 2D EfficientNetv2 architectures across diverse datasets.</p><p><strong>Results: </strong>Model performance was evaluated using the area under the curve (AUC) and the DeLong method for AUC confidence intervals. The models trained on the most diverse datasets showed the highest external testing performance, with AUC values ranging from 0.73 to 0.76 for CT and 0.70 to 0.73 for CXR. Internal testing yielded higher AUC values (0.77 to 0.85 for CT and 0.77 to 1.0 for CXR), highlighting a substantial drop in performance during external validation, which underscores the importance of diverse and comprehensive training and testing data. Most notably, the VIT approach provided an objective assessment of the utility of diverse models and datasets while offering insight into the influence of dataset characteristics, patient factors, and imaging physics on AI efficacy.</p><p><strong>Conclusions: </strong>The VIT approach enhances model transparency and reliability, offering nuanced insights into the factors driving AI performance and bridging the gap between experimental and clinical settings.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014506"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12906637/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146208146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From preoperative computed tomography to postmastoidectomy mesh construction: mastoidectomy shape prediction for cochlear implant surgery. 从术前计算机断层扫描到乳突切除术后网状结构:人工耳蜗手术乳突切除术形状预测。
IF 1.7
Journal of Medical Imaging Pub Date : 2026-01-01 Epub Date: 2026-01-27 DOI: 10.1117/1.JMI.13.1.014004
Yike Zhang, Eduardo Davalos, Dingjie Su, Ange Lou, Jack Noble
{"title":"From preoperative computed tomography to postmastoidectomy mesh construction: mastoidectomy shape prediction for cochlear implant surgery.","authors":"Yike Zhang, Eduardo Davalos, Dingjie Su, Ange Lou, Jack Noble","doi":"10.1117/1.JMI.13.1.014004","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014004","url":null,"abstract":"<p><strong>Purpose: </strong>Cochlear implant (CI) surgery treats severe hearing loss by inserting an electrode array into the cochlea to stimulate the auditory nerve. An important step in this procedure is mastoidectomy, which removes part of the mastoid region of the temporal bone to provide surgical access. Accurate mastoidectomy shape prediction from preoperative imaging improves presurgical planning, reduces risks, and enhances surgical outcomes. Despite its importance, there are limited deep-learning-based studies regarding this topic due to the challenges of acquiring ground-truth labels. We address this gap by investigating self-supervised and weakly-supervised learning models to predict the mastoidectomy region without human annotations.</p><p><strong>Approach: </strong>We propose a hybrid self-supervised and weakly-supervised learning framework to predict the mastoidectomy region directly from preoperative CT scans, where the mastoid remains intact. Our self-supervised learning approach reconstructs the postmastoidectomy 3D surface from preoperative imaging, aiming to align with the corresponding intraoperative microscope views for future surgical navigation-related applications. Postoperative CT scans are used in the self-supervised learning model to assist training procedures despite additional challenges such as metal artifacts and low signal-to-noise ratios introduced by them. To further improve the accuracy and robustness, we introduce a Mamba-based weakly-supervised model that refines mastoidectomy shape prediction by using 3D T-distribution loss function, inspired by the student- <math><mrow><mi>t</mi></mrow> </math> distribution. Weak supervision is achieved by leveraging segmentation results from the prior self-supervised framework, eliminating the manual data labeling process.</p><p><strong>Results: </strong>Our hybrid method achieves a mean Dice score of 0.72 when predicting the complex and boundary-less mastoidectomy shape, surpassing state-of-the-art approaches and demonstrating strong performance. The method provides groundwork for constructing 3D postmastoidectomy surfaces directly from the corresponding preoperative CT scans.</p><p><strong>Conclusion: </strong>To our knowledge, this is the first work that integrates self-supervised and weakly-supervised learning for mastoidectomy shape prediction, offering a robust and efficient solution for CI surgical planning while leveraging 3D T-distribution loss in weakly-supervised medical imaging.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014004"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12838396/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146094516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep learning-based mediastinal lymph node assessment on PET/CT images without pixel-level annotations. 无像素级注释的PET/CT纵隔淋巴结深度学习评估
IF 1.7
Journal of Medical Imaging Pub Date : 2026-01-01 Epub Date: 2026-02-18 DOI: 10.1117/1.JMI.13.1.014507
Sofija Engelson, Yannic Elser, Malte Maria Sieren, Jan Ehrhardt, Julia Andresen, Stefanie Schierholz, Tobias Keck, Daniel Drömann, Jörg Barkhausen, Heinz Handels
{"title":"Deep learning-based mediastinal lymph node assessment on PET/CT images without pixel-level annotations.","authors":"Sofija Engelson, Yannic Elser, Malte Maria Sieren, Jan Ehrhardt, Julia Andresen, Stefanie Schierholz, Tobias Keck, Daniel Drömann, Jörg Barkhausen, Heinz Handels","doi":"10.1117/1.JMI.13.1.014507","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014507","url":null,"abstract":"<p><strong>Purpose: </strong><math><mrow><mi>N</mi></mrow> </math> -staging, a critical component in cancer diagnostics, quantifies metastatic involvement of lymph nodes and plays an important role in guiding treatment decisions. Manual assessment of lymph nodes on PET/CT scans is time-consuming due to minimal contrast to surrounding tissue and strong heterogeneity of the lymph node's morphology. To streamline the <math><mrow><mi>N</mi></mrow> </math> -staging process, we propose a deep learning-based algorithm that localizes lymph node stations through atlas-to-patient registration, classifies mediastinal lymph node stations as malignant or benign, and subsequently performs automated <math><mrow><mi>N</mi></mrow> </math> -staging. Notably, our model is trained without any pixel-level annotations, i.e., using image-level classification labels only.</p><p><strong>Approach: </strong>To address the challenge of training without annotations at the pixel level, we use prior knowledge of the lymph node station locations through atlas-to-patient registration and deduce pseudo-labels for lymph node station groups from the <math><mrow><mi>N</mi></mrow> </math> -stage to enable weakly supervised network training.</p><p><strong>Results: </strong>The proposed algorithm achieves an accuracy of <math><mrow><mn>0.88</mn> <mo>±</mo> <mn>0.02</mn></mrow> </math> , a sensitivity of <math><mrow><mn>0.72</mn> <mo>±</mo> <mn>0.08</mn></mrow> </math> , and a specificity of <math><mrow><mn>0.90</mn> <mo>±</mo> <mn>0.03</mn></mrow> </math> for lymph node station classification, which is significantly better than the performance of the standard threshold-based approach used for lymph node assessment in radiological images and an algorithm for PET lesion segmentation that was trained with segmentation masks. For automatic <math><mrow><mi>N</mi></mrow> </math> -staging, the accuracy of <math><mrow><mn>0.63</mn> <mo>±</mo> <mn>0.04</mn></mrow> </math> is on par with an algorithm that was trained with segmentation masks.</p><p><strong>Conclusions: </strong>The division of the problem setting into subtasks as well as the integration of prior knowledge enables better or comparable performance of models trained with and without segmentation masks.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014507"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12915757/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146229382","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward robust modeling of breast biomechanical compression: an extended study using graph neural networks. 乳房生物力学压缩的鲁棒建模:使用图神经网络的扩展研究。
IF 1.7
Journal of Medical Imaging Pub Date : 2026-01-01 Epub Date: 2025-12-29 DOI: 10.1117/1.JMI.13.1.015001
Hadeel Awwad, Eloy García, Robert Martí
{"title":"Toward robust modeling of breast biomechanical compression: an extended study using graph neural networks.","authors":"Hadeel Awwad, Eloy García, Robert Martí","doi":"10.1117/1.JMI.13.1.015001","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.015001","url":null,"abstract":"<p><strong>Purpose: </strong>Accurate simulation of breast tissue deformation is essential for reliable image registration between 3D imaging modalities and 2D mammograms, where compression significantly alters tissue geometry. Although finite element analysis (FEA) provides high-fidelity modeling, it is computationally intensive and not well suited for rapid simulations. To address this, the physics-based graph neural network (PhysGNN) has been introduced as a computationally efficient approximation model trained on FEA-generated deformations. We extend prior work by evaluating the performance of PhysGNN on new digital breast phantoms and assessing the impact of training on multiple phantoms.</p><p><strong>Approach: </strong>PhysGNN was trained on both single-phantom (per-geometry) and multiphantom (multigeometry) datasets generated from incremental FEA simulations. The digital breast phantoms represent the uncompressed state, serving as input geometries for predicting compressed configurations. A leave-one-deformation-out evaluation strategy was used to assess predictive performance under compression.</p><p><strong>Results: </strong>Training on new digital phantoms confirmed the model's robust performance, though with some variability in prediction accuracy reflecting the diverse anatomical structures. Multiphantom training further enhanced this robustness and reduced prediction errors.</p><p><strong>Conclusions: </strong>PhysGNN offers a computationally efficient alternative to FEA for simulating breast compression. The results showed that model performance remains robust when trained per-geometry, and further demonstrated that multigeometry training enhances predictive accuracy and robustness for the geometries included in the training set. This suggests a strong potential path toward developing reliable models for generating compressed breast volumes, which could facilitate image registration and algorithm development.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"015001"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12745487/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145866013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Attention-driven framework to segment renal ablation zone in posttreatment CT images: a step toward ablation margin evaluation. 治疗后CT图像中肾消融区分割的注意力驱动框架:消融边缘评估的一个步骤。
IF 1.7
Journal of Medical Imaging Pub Date : 2026-01-01 Epub Date: 2026-01-05 DOI: 10.1117/1.JMI.13.1.014001
Maryam Rastegarpoor, Derek W Cool, Aaron Fenster
{"title":"Attention-driven framework to segment renal ablation zone in posttreatment CT images: a step toward ablation margin evaluation.","authors":"Maryam Rastegarpoor, Derek W Cool, Aaron Fenster","doi":"10.1117/1.JMI.13.1.014001","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014001","url":null,"abstract":"<p><strong>Purpose: </strong>Thermal ablation is a minimally invasive therapy used for the treatment of small renal cell carcinoma tumors. Treatment success is evaluated on postablation computed tomography (CT) to determine if the ablation zone covered the tumor with an adequate treatment margin (often 5 to 10 mm). Incorrect margin identification can lead to treatment misassessment, resulting in unnecessary additional ablation. Therefore, segmentation of the renal ablation zone (RAZ) is crucial for treatment evaluation. We aim to develop and assess an accurate deep learning workflow for delineating the RAZ from surrounding tissues in kidney CT images.</p><p><strong>Approach: </strong>We present an advanced deep learning method using the attention-based U-Net architecture to segment the RAZ. The workflow leverages the strengths of U-Net, enhanced with attention mechanisms, to improve the network's focus on the most relevant parts of the images, resulting in an accurate segmentation.</p><p><strong>Results: </strong>Our model was trained and evaluated on a dataset comprising 76 patients' annotated RAZs in CT images. Analysis demonstrated that the proposed workflow achieved an accuracy <math><mrow><mo>=</mo> <mn>0.97</mn> <mo>±</mo> <mn>0.02</mn></mrow> </math> , precision <math><mrow><mo>=</mo> <mn>0.74</mn> <mo>±</mo> <mn>0.23</mn></mrow> </math> , <math><mrow><mtext>recall</mtext> <mo>=</mo> <mn>0.73</mn> <mo>±</mo> <mn>0.25</mn></mrow> </math> , <math><mrow><mi>DSC</mi> <mo>=</mo> <mn>0.70</mn> <mo>±</mo> <mn>0.22</mn></mrow> </math> , Jaccard <math><mrow><mo>=</mo> <mn>0.58</mn> <mo>±</mo> <mn>0.22</mn></mrow> </math> , specificity <math><mrow><mo>=</mo> <mn>0.99</mn> <mo>±</mo> <mn>0.01</mn></mrow> </math> , Hausdorff distance <math><mrow><mo>=</mo> <mn>6.70</mn> <mo>±</mo> <mn>4.44</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> , and mean absolute boundary distance <math><mrow><mo>=</mo> <mn>2.67</mn> <mo>±</mo> <mn>2.22</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> .</p><p><strong>Conclusions: </strong>We used 3D CT images with RAZs and, for the first time, addressed deep-learning-based RAZ segmentation using parallel CT images. Our framework can effectively segment RAZs, allowing clinicians to automatically determine the ablation margin, making our tool ready for clinical use. Prediction time is <math><mrow><mo>∼</mo> <mn>1</mn> <mtext>  </mtext> <mi>s</mi></mrow> </math> per patient, enabling clinicians to perform quick reviews, especially in time-constrained settings.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014001"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12767621/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145913125","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Airway quantifications of bronchitis patients with photon-counting and energy-integrating computed tomography. 光子计数和能量积分计算机断层扫描对支气管炎患者气道的定量研究。
IF 1.7
Journal of Medical Imaging Pub Date : 2026-01-01 Epub Date: 2026-02-02 DOI: 10.1117/1.JMI.13.1.013501
Fong Chi Ho, William Paul Segars, Ehsan Samei, Ehsan Abadi
{"title":"Airway quantifications of bronchitis patients with photon-counting and energy-integrating computed tomography.","authors":"Fong Chi Ho, William Paul Segars, Ehsan Samei, Ehsan Abadi","doi":"10.1117/1.JMI.13.1.013501","DOIUrl":"10.1117/1.JMI.13.1.013501","url":null,"abstract":"<p><strong>Purpose: </strong>Accurate airway measurement is critical for bronchitis quantification with computed tomography (CT), yet optimal protocols and the added value of photon-counting CT (PCCT) over energy-integrating CT (EICT) for reducing bias remain unclear. We quantified biomarker accuracy across modalities and protocols and assessed strategies to reduce bias.</p><p><strong>Approach: </strong>A virtual imaging trial with 20 bronchitis anthropomorphic models was scanned using a validated simulator for two systems (EICT: SOMATOM Flash; PCCT: NAEOTOM Alpha) at 6.3 and 12.6 mGy. Reconstructions varied algorithm, kernel sharpness, slice thickness, and pixel size. Pi10 (square-root wall thickness at 10-mm perimeter) and WA% (wall-area percentage) were compared against ground-truth airway dimensions obtained from the 0.1-mm-precision anatomical models prior to CT simulation. External validation used clinical PCCT ( <math><mrow><mi>n</mi> <mo>=</mo> <mn>22</mn></mrow> </math> ) and EICT ( <math><mrow><mi>n</mi> <mo>=</mo> <mn>80</mn></mrow> </math> ).</p><p><strong>Results: </strong>Simulated airway dimensions agreed with pathological references ( <math><mrow><mi>R</mi> <mo>=</mo> <mn>0.89</mn> <mo>-</mo> <mn>0.93</mn></mrow> </math> ). PCCT had lower errors than EICT across segmented generations ( <math><mrow><mi>p</mi> <mo><</mo> <mn>0.05</mn></mrow> </math> ). Under optimal parameters, PCCT improved Pi10 and WA% accuracy by 26.3% and 64.9%. Across the tested PCCT and EICT imaging protocols, improvements were associated with sharper kernels (25.8% Pi10, 33.0% WA%), thinner slices (23.9% Pi10, 49.8% WA%), smaller pixels (17.0% Pi10, 23.1% WA%), and higher dose ( <math><mrow><mo>≤</mo> <mn>3.9</mn> <mo>%</mo></mrow> </math> ). Clinically, PCCT achieved higher maximum airway generation ( <math><mrow><mn>8.8</mn> <mo>±</mo> <mn>0.5</mn></mrow> </math> versus <math><mrow><mn>6.0</mn> <mo>±</mo> <mn>1.1</mn></mrow> </math> ) and lower variability, mirroring trends in virtual results.</p><p><strong>Conclusions: </strong>PCCT improves the accuracy and consistency of airway biomarker quantification relative to EICT, particularly with optimized protocols. The validated virtual platform enables modality-bias assessment and protocol optimization for accurate, reproducible bronchitis measurements.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"013501"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12863983/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146114479","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LCSD-Net: a light-weight cross-attention-based semantic dual transformer for domain generalization in melanoma detection. LCSD-Net:用于黑色素瘤检测领域泛化的轻量级交叉注意语义双转换器。
IF 1.7
Journal of Medical Imaging Pub Date : 2026-01-01 Epub Date: 2026-01-06 DOI: 10.1117/1.JMI.13.1.014502
Rishi Agrawal, Neeraj Gupta, Anand Singh Jalal
{"title":"LCSD-Net: a light-weight cross-attention-based semantic dual transformer for domain generalization in melanoma detection.","authors":"Rishi Agrawal, Neeraj Gupta, Anand Singh Jalal","doi":"10.1117/1.JMI.13.1.014502","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014502","url":null,"abstract":"<p><strong>Purpose: </strong>Research in deep learning has shown a great advancement in the detection of melanoma. However, recent literature has emphasized a tendency of certain models to rely on disease-irrelevant visual artifacts such as dark corners, dense hair, or ruler marks. The dependence on these markers leads to biased models that do well for training but generalize poorly to heterogeneous clinical environments. To address these limitations in developing reliability in skin lesion detection, a lightweight cross-attention-based semantic dual (LCSD) transformer model was proposed.</p><p><strong>Approach: </strong>The LCSD model extracts global-level semantic information, uses feature normalization to improve model accuracy, and employs semantic queries to improve domain generalization. Multihead attention is included with the semantic queries to refine global features. The cross-attention between feature maps and semantic query provides the model with a generalized encoding of the global context. The model improved the computational complexity from <math><mrow><mi>O</mi> <mo>(</mo> <msup><mrow><mi>n</mi></mrow> <mrow><mn>2</mn></mrow> </msup> <mi>d</mi> <mo>)</mo></mrow> </math> to <math><mrow><mi>O</mi> <mo>(</mo> <mi>n</mi> <mi>m</mi> <mi>d</mi> <mo>+</mo> <msup><mrow><mi>m</mi></mrow> <mrow><mn>2</mn></mrow> </msup> <mi>d</mi> <mo>)</mo></mrow> </math> , which makes the model suitable for the development of real-time and mobile applications.</p><p><strong>Results: </strong>Empirical evaluation was conducted on three challenging datasets: Derm7pt-Dermoscopic, Derm7pt-Clinical, and PAD-UFES-20. The proposed model achieved classification accuracies of 82.82%, 72.95%, and 86.21%, respectively. These results demonstrate superior performance compared with conventional transformer-based models, highlighting both improved robustness and reduced computational cost.</p><p><strong>Conclusion: </strong>The LCSD model mitigates the influence of irrelevant visual characteristics, enhances domain generalization, and ensures better adaptability across diverse clinical scenarios. Its lightweight design further supports deployment in mobile applications, making it a reliable and efficient solution for real-world melanoma detection.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014502"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12773922/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145918799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信
小红书