{"title":"Reliable Delineation of Clinical Target Volumes for Cervical Cancer Radiotherapy on CT/MR Dual-Modality Images","authors":"Ying Sun, Yuening Wang, Kexin Gan, Yuxin Wang, Ying Chen, Yun Ge, Jie Yuan, Hanzi Xu","doi":"10.1007/s10278-023-00951-5","DOIUrl":"https://doi.org/10.1007/s10278-023-00951-5","url":null,"abstract":"<p>Accurate delineation of the clinical target volume (CTV) is a crucial prerequisite for safe and effective radiotherapy characterized. This study addresses the integration of magnetic resonance (MR) images to aid in target delineation on computed tomography (CT) images. However, obtaining MR images directly can be challenging. Therefore, we employ AI-based image generation techniques to “intelligentially generate” MR images from CT images to improve CTV delineation based on CT images. To generate high-quality MR images, we propose an attention-guided single-loop image generation model. The model can yield higher-quality images by introducing an attention mechanism in feature extraction and enhancing the loss function. Based on the generated MR images, we propose a CTV segmentation model fusing multi-scale features through image fusion and a hollow space pyramid module to enhance segmentation accuracy. The image generation model used in this study improves the peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) from 14.87 and 0.58 to 16.72 and 0.67, respectively, and improves the feature distribution distance and learning-perception image similarity from 180.86 and 0.28 to 110.98 and 0.22, achieving higher quality image generation. The proposed segmentation method demonstrates high accuracy, compared with the FCN method, the intersection over union ratio and the Dice coefficient are improved from 0.8360 and 0.8998 to 0.9043 and 0.9473, respectively. Hausdorff distance and mean surface distance decreased from 5.5573 mm and 2.3269 mm to 4.7204 mm and 0.9397 mm, respectively, achieving clinically acceptable segmentation accuracy. Our method might reduce physicians’ manual workload and accelerate the diagnosis and treatment process while decreasing inter-observer variability in identifying anatomical structures.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"1 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yun Wang, Deng Lyu, Lei Hu, Junhong Wu, Shaofeng Duan, Taohu Zhou, Wenting Tu, Yi Xiao, Li Fan, Shiyuan Liu
{"title":"CT-Based Intratumoral and Peritumoral Radiomics Nomograms for the Preoperative Prediction of Spread Through Air Spaces in Clinical Stage IA Non-small Cell Lung Cancer","authors":"Yun Wang, Deng Lyu, Lei Hu, Junhong Wu, Shaofeng Duan, Taohu Zhou, Wenting Tu, Yi Xiao, Li Fan, Shiyuan Liu","doi":"10.1007/s10278-023-00939-1","DOIUrl":"https://doi.org/10.1007/s10278-023-00939-1","url":null,"abstract":"<p>The study aims to investigate the value of intratumoral and peritumoral radiomics and clinical-radiological features for predicting spread through air spaces (STAS) in patients with clinical stage IA non-small cell lung cancer (NSCLC). A total of 336 NSCLC patients from our hospital were randomly divided into the training cohort (<i>n</i> = 236) and the internal validation cohort (<i>n</i> = 100) at a ratio of 7:3, and 69 patients from the other two external hospitals were collected as the external validation cohort. Univariate and multivariate analyses were used to select clinical-radiological features and construct a clinical model. The GTV, PTV5, PTV10, PTV15, PTV20, GPTV5, GPTV10, GPTV15, and GPTV20 models were constructed based on intratumoral and peritumoral (5 mm, 10 mm, 15 mm, 20 mm) radiomics features. Additionally, the radscore of the optimal radiomics model and clinical-radiological predictors were used to construct a combined model and plot a nomogram. Lastly, the ROC curve and AUC value were used to evaluate the diagnostic performance of the model. Tumor density type (OR = 6.738) and distal ribbon sign (OR = 5.141) were independent risk factors for the occurrence of STAS. The GPTV10 model outperformed the other radiomics models, and its AUC values were 0.887, 0.876, and 0.868 in the three cohorts. The AUC values of the combined model constructed based on GPTV10 radscore and clinical-radiological predictors were 0.901, 0.875, and 0.878. DeLong test results revealed that the combined model was superior to the clinical model in the three cohorts. The nomogram based on GPTV10 radscore and clinical-radiological features exhibited high predictive efficiency for STAS status in NSCLC.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"24 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421588","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Deep Learning–based Diagnosis of Pulmonary Tuberculosis on Chest X-ray in the Emergency Department: A Retrospective Study","authors":"Chih-Hung Wang, Weishan Chang, Meng-Rui Lee, Joyce Tay, Cheng-Yi Wu, Meng-Che Wu, Holger R. Roth, Dong Yang, Can Zhao, Weichung Wang, Chien-Hua Huang","doi":"10.1007/s10278-023-00952-4","DOIUrl":"https://doi.org/10.1007/s10278-023-00952-4","url":null,"abstract":"<p>Prompt and correct detection of pulmonary tuberculosis (PTB) is critical in preventing its spread. We aimed to develop a deep learning–based algorithm for detecting PTB on chest X-ray (CXRs) in the emergency department. This retrospective study included 3498 CXRs acquired from the National Taiwan University Hospital (NTUH). The images were chronologically split into a training dataset, NTUH-1519 (images acquired during the years 2015 to 2019; <i>n</i> = 2144), and a testing dataset, NTUH-20 (images acquired during the year 2020; <i>n</i> = 1354). Public databases, including the NIH ChestX-ray14 dataset (model training; 112,120 images), Montgomery County (model testing; 138 images), and Shenzhen (model testing; 662 images), were also used in model development. EfficientNetV2 was the basic architecture of the algorithm. Images from ChestX-ray14 were employed for pseudo-labelling to perform semi-supervised learning. The algorithm demonstrated excellent performance in detecting PTB (area under the receiver operating characteristic curve [AUC] 0.878, 95% confidence interval [CI] 0.854–0.900) in NTUH-20. The algorithm showed significantly better performance in posterior-anterior (PA) CXR (AUC 0.940, 95% CI 0.912–0.965, <i>p-value</i> < 0.001) compared with anterior–posterior (AUC 0.782, 95% CI 0.644–0.897) or portable anterior–posterior (AUC 0.869, 95% CI 0.814–0.918) CXR. The algorithm accurately detected cases of bacteriologically confirmed PTB (AUC 0.854, 95% CI 0.823–0.883). Finally, the algorithm tested favourably in Montgomery County (AUC 0.838, 95% CI 0.765–0.904) and Shenzhen (AUC 0.806, 95% CI 0.771–0.839). A deep learning–based algorithm could detect PTB on CXR with excellent performance, which may help shorten the interval between detection and airborne isolation for patients with PTB.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"39 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139423453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust Medical Diagnosis: A Novel Two-Phase Deep Learning Framework for Adversarial Proof Disease Detection in Radiology Images","authors":"Sheikh Burhan ul haque, Aasim Zafar","doi":"10.1007/s10278-023-00916-8","DOIUrl":"https://doi.org/10.1007/s10278-023-00916-8","url":null,"abstract":"<p>In the realm of medical diagnostics, the utilization of deep learning techniques, notably in the context of radiology images, has emerged as a transformative force. The significance of artificial intelligence (AI), specifically machine learning (ML) and deep learning (DL), lies in their capacity to rapidly and accurately diagnose diseases from radiology images. This capability has been particularly vital during the COVID-19 pandemic, where rapid and precise diagnosis played a pivotal role in managing the spread of the virus. DL models, trained on vast datasets of radiology images, have showcased remarkable proficiency in distinguishing between normal and COVID-19-affected cases, offering a ray of hope amidst the crisis. However, as with any technological advancement, vulnerabilities emerge. Deep learning-based diagnostic models, although proficient, are not immune to adversarial attacks. These attacks, characterized by carefully crafted perturbations to input data, can potentially disrupt the models’ decision-making processes. In the medical context, such vulnerabilities could have dire consequences, leading to misdiagnoses and compromised patient care. To address this, we propose a two-phase defense framework that combines advanced adversarial learning and adversarial image filtering techniques. We use a modified adversarial learning algorithm to enhance the model’s resilience against adversarial examples during the training phase. During the inference phase, we apply JPEG compression to mitigate perturbations that cause misclassification. We evaluate our approach on three models based on ResNet-50, VGG-16, and Inception-V3. These models perform exceptionally in classifying radiology images (X-ray and CT) of lung regions into normal, pneumonia, and COVID-19 pneumonia categories. We then assess the vulnerability of these models to three targeted adversarial attacks: fast gradient sign method (FGSM), projected gradient descent (PGD), and basic iterative method (BIM). The results show a significant drop in model performance after the attacks. However, our defense framework greatly improves the models’ resistance to adversarial attacks, maintaining high accuracy on adversarial examples. Importantly, our framework ensures the reliability of the models in diagnosing COVID-19 from clean images.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"41 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139420755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Automatic Urinary Stone Detection System for Abdominal Non-Enhanced CT Images Reduces the Burden on Radiologists","authors":"Zhaoyu Xing, Zuhui Zhu, Zhenxing Jiang, Jingshi Zhao, Qin Chen, Wei Xing, Liang Pan, Yan Zeng, Aie Liu, Jiule Ding","doi":"10.1007/s10278-023-00946-2","DOIUrl":"https://doi.org/10.1007/s10278-023-00946-2","url":null,"abstract":"<p>To develop a fully automatic urinary stone detection system (kidney, ureter, and bladder) and to test it in a real clinical environment. The local institutional review board approved this retrospective single-center study that used non-enhanced abdominopelvic CT scans from patients admitted urology (uPatients) and emergency (ePatients). The uPatients were randomly divided into training and validation sets in a ratio of 3:1. We designed a cascade urinary stone map location-feature pyramid networks (USm-FPNs) and innovatively proposed a ureter distance heatmap method to estimate the ureter position on non-enhanced CT to further reduce the false positives. The performances of the system were compared using the free-response receiver operating characteristic curve and the precision-recall curve. This study included 811 uPatients and 356 ePatients. At stone level, the cascade detector USm-FPNs has the mean of false positives per scan (mFP) 1.88 with the sensitivity 0.977 in validation set, and mFP was further reduced to 1.18 with the sensitivity 0.977 after combining the ureter distance heatmap. At patient level, the sensitivity and precision were as high as 0.995 and 0.990 in validation set, respectively. In a real clinical set of ePatients (27.5% of patients contain stones), the mFP was 1.31 with as high as sensitivity 0.977, and the diagnostic time reduced by > 20% with the system help. A fully automatic detection system for entire urinary stones on non-enhanced CT scans was proposed and reduces obviously the burden on junior radiologists without compromising sensitivity in real emergency data.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"69 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robustness of Deep Networks for Mammography: Replication Across Public Datasets","authors":"","doi":"10.1007/s10278-023-00943-5","DOIUrl":"https://doi.org/10.1007/s10278-023-00943-5","url":null,"abstract":"<h3>Abstract</h3> <p>Deep neural networks have demonstrated promising performance in screening mammography with recent studies reporting performance at or above the level of trained radiologists on internal datasets. However, it remains unclear whether the performance of these trained models is robust and replicates across external datasets. In this study, we evaluate four state-of-the-art publicly available models using four publicly available mammography datasets (CBIS-DDSM, INbreast, CMMD, OMI-DB). Where test data was available, published results were replicated. The best-performing model, which achieved an area under the ROC curve (AUC) of 0.88 on internal data from NYU, achieved here an AUC of 0.9 on the external CMMD dataset (<em>N</em> = 826 exams). On the larger OMI-DB dataset (<em>N</em> = 11,440 exams), it achieved an AUC of 0.84 but did not match the performance of individual radiologists (at a specificity of 0.92, the sensitivity was 0.97 for the radiologist and 0.53 for the network for a 1-year follow-up). The network showed higher performance for in situ cancers, as opposed to invasive cancers. Among invasive cancers, it was relatively weaker at identifying asymmetries and was relatively stronger at identifying masses. The three other trained models that we evaluated all performed poorly on external datasets. Independent validation of trained models is an essential step to ensure safe and reliable use. Future progress in AI for mammography may depend on a concerted effort to make larger datasets publicly available that span multiple clinical sites.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"2 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Xinhua Dai, Anjie Liu, Junhong Liu, Mengjun Zhan, Yuanyuan Liu, Wenchi Ke, Lei Shi, Xinyu Huang, Hu Chen, Zhenhua Deng, Fei Fan
{"title":"Machine Learning Supported the Modified Gustafson’s Criteria for Dental Age Estimation in Southwest China","authors":"Xinhua Dai, Anjie Liu, Junhong Liu, Mengjun Zhan, Yuanyuan Liu, Wenchi Ke, Lei Shi, Xinyu Huang, Hu Chen, Zhenhua Deng, Fei Fan","doi":"10.1007/s10278-023-00956-0","DOIUrl":"https://doi.org/10.1007/s10278-023-00956-0","url":null,"abstract":"<p>Adult age estimation is one of the most challenging problems in forensic science and physical anthropology. In this study, we aimed to develop and evaluate machine learning (ML) methods based on the modified Gustafson’s criteria for dental age estimation. In this retrospective study, a total of 851 orthopantomograms were collected from patients aged 15 to 40 years old. The secondary dentin formation (SE), periodontal recession (PE), and attrition (AT) of four mandibular premolars were analyzed according to the modified Gustafson’s criteria. Ten ML models were generated and compared for age estimation. The partial least squares regressor outperformed other models in males with a mean absolute error (MAE) of 4.151 years. The support vector regressor (MAE = 3.806 years) showed good performance in females. The accuracy of ML models is better than the single-tooth model provided in the previous studies (MAE = 4.747 years in males and MAE = 4.957 years in females). The Shapley additive explanations method was used to reveal the importance of the 12 features in ML models and found that AT and PE are the most influential in age estimation. The findings suggest that the modified Gustafson method can be effectively employed for adult age estimation in the southwest Chinese population. Furthermore, this study highlights the potential of machine learning models to assist experts in achieving accurate and interpretable age estimation.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"118 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ling Ma, Gen Li, Xingyu Feng, Qiliang Fan, Lizhi Liu
{"title":"TiCNet: Transformer in Convolutional Neural Network for Pulmonary Nodule Detection on CT Images","authors":"Ling Ma, Gen Li, Xingyu Feng, Qiliang Fan, Lizhi Liu","doi":"10.1007/s10278-023-00904-y","DOIUrl":"https://doi.org/10.1007/s10278-023-00904-y","url":null,"abstract":"<p>Lung cancer is the leading cause of cancer death. Since lung cancer appears as nodules in the early stage, detecting the pulmonary nodules in an early phase could enhance the treatment efficiency and improve the survival rate of patients. The development of computer-aided analysis technology has made it possible to automatically detect lung nodules in Computed Tomography (CT) screening. In this paper, we propose a novel detection network, TiCNet. It is attempted to embed a transformer module in the 3D Convolutional Neural Network (CNN) for pulmonary nodule detection on CT images. First, we integrate the transformer and CNN in an end-to-end structure to capture both the short- and long-range dependency to provide rich information on the characteristics of nodules. Second, we design the attention block and multi-scale skip pathways for improving the detection of small nodules. Last, we develop a two-head detector to guarantee high sensitivity and specificity. Experimental results on the LUNA16 dataset and PN9 dataset showed that our proposed TiCNet achieved superior performance compared with existing lung nodule detection methods. Moreover, the effectiveness of each module has been proven. The proposed TiCNet model is an effective tool for pulmonary nodule detection. Validation revealed that this model exhibited excellent performance, suggesting its potential usefulness to support lung cancer screening.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"85 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Chongfeng Duan, Dapeng Hao, Jiufa Cui, Gang Wang, Wenjian Xu, Nan Li, Xuejun Liu
{"title":"An MRI-Based Deep Transfer Learning Radiomics Nomogram to Predict Ki-67 Proliferation Index of Meningioma","authors":"Chongfeng Duan, Dapeng Hao, Jiufa Cui, Gang Wang, Wenjian Xu, Nan Li, Xuejun Liu","doi":"10.1007/s10278-023-00937-3","DOIUrl":"https://doi.org/10.1007/s10278-023-00937-3","url":null,"abstract":"<p>The objective of this study was to predict Ki-67 proliferation index of meningioma by using a nomogram based on clinical, radiomics, and deep transfer learning (DTL) features. A total of 318 cases were enrolled in the study. The clinical, radiomics, and DTL features were selected to construct models. The calculation of radiomics and DTL score was completed by using selected features and correlation coefficient. The deep transfer learning radiomics (DTLR) nomogram was constructed by selected clinical features, radiomics score, and DTL score. The area under the receiver operator characteristic curve (AUC) was calculated. The models were compared by Delong test of AUCs and decision curve analysis (DCA). The features of sex, size, and peritumoral edema were selected to construct clinical model. Seven radiomics features and 15 DTL features were selected. The AUCs of clinical, radiomics, DTL model, and DTLR nomogram were 0.746, 0.75, 0.717, and 0.779 respectively. DTLR nomogram had the highest AUC of 0.779 (95% CI 0.6643–0.8943) with an accuracy rate of 0.734, a sensitivity value of 0.719, and a specificity value of 0.75 in test set. There was no significant difference in AUCs among four models in Delong test. The DTLR nomogram had a larger net benefit than other models across all the threshold probability. The DTLR nomogram had a satisfactory performance in Ki-67 prediction and could be a new evaluation method of meningioma which would be useful in the clinical decision-making.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"40 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"PET KinetiX—A Software Solution for PET Parametric Imaging at the Whole Field of View Level","authors":"Florent L. Besson, Sylvain Faure","doi":"10.1007/s10278-023-00965-z","DOIUrl":"https://doi.org/10.1007/s10278-023-00965-z","url":null,"abstract":"<p>Kinetic modeling represents the ultimate foundations of PET quantitative imaging, a unique opportunity to better characterize the diseases or prevent the reduction of drugs development. Primarily designed for research, parametric imaging based on PET kinetic modeling may become a reality in future clinical practice, enhanced by the technical abilities of the latest generation of commercially available PET systems. In the era of precision medicine, such paradigm shift should be promoted, regardless of the PET system. In order to anticipate and stimulate this emerging clinical paradigm shift, we developed a constructor-independent software package, called PET KinetiX, allowing a faster and easier computation of parametric images from any 4D PET DICOM series, at the whole field of view level. The PET KinetiX package is currently a plug-in for Osirix DICOM viewer. The package provides a suite of five PET kinetic models: Patlak, Logan, 1-tissue compartment model, 2-tissue compartment model, and first pass blood flow. After uploading the 4D-PET DICOM series into Osirix, the image processing requires very few steps: the choice of the kinetic model and the definition of an input function. After a 2-min process, the PET parametric and error maps of the chosen model are automatically estimated voxel-wise and written in DICOM format. The software benefits from the graphical user interface of Osirix, making it user-friendly. Compared to PMOD-PKIN (version 4.4) on twelve <sup>18</sup>F-FDG PET dynamic datasets, PET KinetiX provided an absolute bias of 0.1% (0.05–0.25) and 5.8% (3.3–12.3) for Ki<sub>Patlak</sub> and Ki<sub>2TCM</sub>, respectively. Several clinical research illustrative cases acquired on different hybrid PET systems (standard or extended axial fields of view, PET/CT, and PET/MRI), with different acquisition schemes (single-bed single-pass or multi-bed multipass), are also provided. PET KinetiX is a very fast and efficient independent research software that helps molecular imaging users easily and quickly produce 3D PET parametric images from any reconstructed 4D-PET data acquired on standard or large PET systems.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"62 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}