Nancy R Newlin, Michael E Kim, Praitayini Kanakaraj, Kimberly Pechman, Niranjana Shashikumar, Elizabeth Moore, Derek Archer, Timothy Hohman, Angela Jefferson, Daniel Moyer, Bennett A Landman
{"title":"Learning disentangled representations to harmonize connectome network measures.","authors":"Nancy R Newlin, Michael E Kim, Praitayini Kanakaraj, Kimberly Pechman, Niranjana Shashikumar, Elizabeth Moore, Derek Archer, Timothy Hohman, Angela Jefferson, Daniel Moyer, Bennett A Landman","doi":"10.1117/1.JMI.12.1.014004","DOIUrl":"10.1117/1.JMI.12.1.014004","url":null,"abstract":"<p><strong>Purpose: </strong>Connectome network metrics are commonly regarded as fundamental properties of the brain, and their alterations have been implicated in the development of Alzheimer's disease, multiple sclerosis, and traumatic brain injury. However, these metrics are actually estimated properties through a multistage propagation from local voxel diffusion estimations, regional tractography, and region of interest mapping. These estimation processes are significantly influenced by choices specific to imaging protocols and software, producing site-wise effects.</p><p><strong>Approach: </strong>Recent advances in disentanglement techniques offer opportunities to learn representational spaces that separate factors that cause domain shifts from intrinsic biological factors. Although these techniques have been applied in unsupervised brain anomaly detection and image-level features, their application to the unique manifold structures of connectome adjacency matrices remains unexplored. Here, we explore the conditional variational autoencoder structure for generating site-invariant representations of the connectome, allowing the harmonization of brain network measures.</p><p><strong>Results: </strong>Focusing on the context of aging, we conducted a study involving 823 patients across two sites. This approach effectively segregates site-specific influences from biological features, aligns network measures across different domains (Cohen's <math><mrow><mi>D</mi> <mo><</mo> <mn>0.2</mn></mrow> </math> and Mann-Whitney <math><mrow><mi>U</mi> <mtext>-</mtext> <mtext>test</mtext> <mo><</mo> <mn>0.05</mn></mrow> </math> ), and maintains associations with age ( <math><mrow><mn>2.71</mn> <mo>×</mo> <msup><mrow><mn>10</mn></mrow> <mrow><mo>-</mo> <mn>02</mn></mrow> </msup> <mo>±</mo> <mn>2.86</mn> <mo>×</mo> <msup><mrow><mn>10</mn></mrow> <mrow><mo>-</mo> <mn>03</mn></mrow> </msup> </mrow> </math> error in years) and sex ( <math><mrow><mn>0.92</mn> <mo>±</mo> <mn>0.02</mn></mrow> </math> accuracy).</p><p><strong>Conclusions: </strong>Our findings demonstrate that using latent representations significantly harmonizes network measures and provides robust metrics for multi-site brain network analysis.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 1","pages":"014004"},"PeriodicalIF":1.9,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11826349/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143434157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Andrey Makeev, Kaiyan Li, Mark A Anastasio, Arthur Emig, Paul Jahnke, Stephen J Glick
{"title":"Automated assessment of task-based performance of digital mammography and tomosynthesis systems using an anthropomorphic breast phantom and deep learning-based scoring.","authors":"Andrey Makeev, Kaiyan Li, Mark A Anastasio, Arthur Emig, Paul Jahnke, Stephen J Glick","doi":"10.1117/1.JMI.12.S1.S13005","DOIUrl":"10.1117/1.JMI.12.S1.S13005","url":null,"abstract":"<p><strong>Purpose: </strong>Conventional metrics used for assessing digital mammography (DM) and digital breast tomosynthesis (DBT) image quality, including noise, spatial resolution, and detective quantum efficiency, do not necessarily predict how well the system will perform in a clinical task. A number of existing phantom-based methods have their own limitations, such as unrealistic uniform backgrounds, subjective scoring using humans, and regular signal patterns unrepresentative of common clinical findings. We attempted to address this problem with a realistic breast phantom with random hydroxyapatite microcalcifications and semi-automated deep learning-based image scoring. Our goal was to develop a methodology for objective task-based assessment of image quality for tomosynthesis and DM systems, which includes an anthropomorphic phantom, a detection task (microcalcification clusters), and automated performance evaluation using a convolutional neural network.</p><p><strong>Approach: </strong>Experimental 2D and pseudo-3D mammograms of an anthropomorphic inkjet-printed breast phantom with inserted microcalcification clusters were collected on clinical mammography systems to train a signal-present/signal-absent image classifier based on Resnet-18 architecture. In a separate validation study using simulations, this Resnet-18 classifier was shown to approach the performance of an ideal observer. Microcalcification detection performance was evaluated as a function of four dose levels using receiver operating characteristic (ROC) analysis [i.e., area under the ROC curve (AUC)]. To demonstrate the use of this evaluation approach for assessing different technologies, the method was applied to two different mammography systems, as well as to mammograms with re-binned pixels emulating a lower-resolution X-ray detector.</p><p><strong>Results: </strong>Microcalcification detectability, as assessed by the deep learning classifier, was observed to vary with the exposure incident on the breast phantom for both DM and tomosynthesis. At full dose, experimental AUC was 0.96 (for DM) and 0.95 (for DBT), whereas at half dose, it dropped to 0.85 and 0.71, respectively. AUC performance on DM was significantly decreased with an effective larger pixel size obtained with re-binning. The task-based assessment approach also showed the superiority of a newer mammography system compared with an older system.</p><p><strong>Conclusions: </strong>An objective task-based methodology for assessing the image quality of mammography and tomosynthesis systems is proposed. Possible uses for this tool could be quality control, acceptance, and constancy testing, assessing the safety and effectiveness of new technology for regulatory submissions, and system optimization. The results from this study showed that the proposed evaluation method using a deep learning model observer can track differences in microcalcification signal detectability with varied exposure conditions.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 Suppl 1","pages":"S13005"},"PeriodicalIF":1.9,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11474246/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142477762","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Predicting the risk of type 2 diabetes mellitus (T2DM) emergence in 5 years using mammography images: a comparison study between radiomics and deep learning algorithm.","authors":"Nishta Letchumanan, Shouhei Hanaoka, Tomomi Takenaga, Yusuke Suzuki, Takahiro Nakao, Yukihiro Nomura, Takeharu Yoshikawa, Osamu Abe","doi":"10.1117/1.JMI.12.1.014501","DOIUrl":"https://doi.org/10.1117/1.JMI.12.1.014501","url":null,"abstract":"<p><strong>Purpose: </strong>The prevalence of type 2 diabetes mellitus (T2DM) has been steadily increasing over the years. We aim to predict the occurrence of T2DM using mammography images within 5 years using two different methods and compare their performance.</p><p><strong>Approach: </strong>We examined 312 samples, including 110 positive cases (developed T2DM after 5 years) and 202 negative cases (did not develop T2DM) using two different methods. In the first method, a radiomics-based approach, we utilized radiomics features and machine learning (ML) algorithms. The entire breast region was chosen as the region of interest for extracting radiomics features. Then, a binary breast image was created from which we extracted 668 features and analyzed them using various ML algorithms. In the second method, a complex convolutional neural network (CNN) with a modified ResNet architecture and various kernel sizes was applied to raw mammography images for the prediction task. A nested, stratified five-fold cross-validation was done for both parts A and B to compute accuracy, sensitivity, specificity, and area under the receiver operating curve (AUROC). Hyperparameter tuning was also done to enhance the model's performance and reliability.</p><p><strong>Results: </strong>The radiomics approach's light gradient boosting model gave 68.9% accuracy, 30.7% sensitivity, 89.5% specificity, and 0.63 AUROC. The CNN method achieved an AUROC of 0.58 over 20 epochs.</p><p><strong>Conclusion: </strong>Radiomics outperformed CNN by 0.05 in terms of AUROC. This may be due to the more straightforward interpretability and clinical relevance of predefined radiomics features compared with the complex, abstract features learned by CNNs.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 1","pages":"014501"},"PeriodicalIF":1.9,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11702674/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142956750","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Rhea Carlson, Courtney Comrie, Justina Bonaventura, Kellys Morara, Noelle Daigle, Elizabeth Hutchinson, Travis W Sawyer
{"title":"Backscattering Mueller matrix polarimetry estimates microscale anisotropy and orientation in complex brain tissue structure.","authors":"Rhea Carlson, Courtney Comrie, Justina Bonaventura, Kellys Morara, Noelle Daigle, Elizabeth Hutchinson, Travis W Sawyer","doi":"10.1117/1.JMI.12.1.016001","DOIUrl":"10.1117/1.JMI.12.1.016001","url":null,"abstract":"<p><strong>Purpose: </strong>Diffusion magnetic resonance imaging (dMRI) quantitatively estimates brain microstructure, diffusion tractography being one clinically utilized framework. To advance such dMRI approaches, direct quantitative comparisons between microscale anisotropy and orientation are imperative. Complete backscattering Mueller matrix polarized light imaging (PLI) enables the imaging of thin and thick tissue specimens to acquire numerous optical metrics not possible through conventional transmission PLI methods. By comparing complete PLI to dMRI within the ferret optic chiasm (OC), we may investigate the potential of this PLI technique as a dMRI validation tool and gain insight into the microstructural and orientational sensitivity of this imaging method in different tissue thicknesses.</p><p><strong>Approach: </strong>Post-mortem ferret brain tissue samples (whole brain, <math><mrow><mi>n</mi> <mo>=</mo> <mn>1</mn></mrow> </math> and OC, <math><mrow><mi>n</mi> <mo>=</mo> <mn>3</mn></mrow> </math> ) were imaged with both dMRI and complete backscattering Mueller matrix PLI. The specimens were sectioned and then reimaged with PLI. Region of interest and correlation analyses were performed on scalar metrics and orientation vectors of both dMRI and PLI in the coherent optic nerve and crossing chiasm.</p><p><strong>Results: </strong>Optical retardance and dMRI fractional anisotropy showed similar trends between metric values and were strongly correlated, indicating a bias to macroscale architecture in retardance. Thick tissue displays comparable orientation between the diattenuation angle and dMRI fiber orientation distribution glyphs that are not evident in the retardance angle.</p><p><strong>Conclusions: </strong>We demonstrate that backscattering Mueller matrix PLI shows potential as a tool for microstructural dMRI validation in thick tissue specimens. Performing complete polarimetry can provide directional characterization and potentially microscale anisotropy information not available by conventional PLI alone.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 1","pages":"016001"},"PeriodicalIF":1.9,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11686408/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142915974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Examining the influence of digital phantom models in virtual imaging trials for tomographic breast imaging.","authors":"Amar Kavuri, Mini Das","doi":"10.1117/1.JMI.12.1.015501","DOIUrl":"10.1117/1.JMI.12.1.015501","url":null,"abstract":"<p><strong>Purpose: </strong>Digital phantoms are one of the key components of virtual imaging trials (VITs) that aim to assess and optimize new medical imaging systems and algorithms. However, these phantoms vary in their voxel resolution, appearance, and structural details. We investigate whether and how variations between digital phantoms influence system optimization with digital breast tomosynthesis (DBT) as a chosen modality.</p><p><strong>Methods: </strong>We selected widely used and open-access digital breast phantoms created with different methods and generated an ensemble of DBT images to test acquisition strategies. Human observer performance was evaluated using localization receiver operating characteristic (LROC) studies for each phantom type. Noise power spectrum and gaze metrics were also employed to compare phantoms and generated images.</p><p><strong>Results: </strong>Our LROC results show that the arc samplings for peak performance were <math><mrow><mo>∼</mo> <mn>2.5</mn> <mtext> </mtext> <mi>deg</mi></mrow> </math> and 6 deg in Bakic and XCAT breast phantoms, respectively, for the 3-mm lesion detection task and indicate that system optimization outcomes from VITs can vary with phantom types and structural frequency components. In addition, a significant correlation ( <math><mrow><mi>p</mi> <mo><</mo> <mn>0.01</mn></mrow> </math> ) between gaze metrics and diagnostic performance suggests that gaze analysis can be used to understand and evaluate task difficulty in VITs.</p><p><strong>Conclusion: </strong>Our results point to the critical need to evaluate realism in digital phantoms and ensure sufficient structural variations at spatial frequencies relevant to the intended task. Standardizing phantom generation and validation tools may help reduce discrepancies among independently conducted VITs for system or algorithmic optimizations.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 1","pages":"015501"},"PeriodicalIF":1.9,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11686409/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142915976","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sriharsha Marupudi, Joseph A Manus, Muhammad U Ghani, Stephen J Glick, Bahaa Ghammraoui
{"title":"Evaluation of charge summing correction in CdTe-based photon-counting detectors for breast CT: performance metrics and image quality.","authors":"Sriharsha Marupudi, Joseph A Manus, Muhammad U Ghani, Stephen J Glick, Bahaa Ghammraoui","doi":"10.1117/1.JMI.12.1.013501","DOIUrl":"10.1117/1.JMI.12.1.013501","url":null,"abstract":"<p><strong>Purpose: </strong>We evaluate the impact of charge summing correction on a cadmium telluride (CdTe)-based photon-counting detector in breast computed tomography (CT).</p><p><strong>Approach: </strong>We employ a custom-built laboratory benchtop system using the X-THOR FX30 0.75-mm CdTe detector (Varex Imaging, Salt Lake City, Utah, United States) with a pixel pitch of 0.1 mm, operated in both standard mode [single pixel (SP)] and charge summing correction mode [anticoincidence (AC)]. A tungsten anode source operated at 55 kVp with 2-mm aluminum external filtration and tube currents of 25, 100, and 200 mA with corresponding exposure times of 20, 5, and 2.5 ms were employed to study the effects of X-ray fluence and pulse pileup. Performance comparisons between AC and SP modes are performed in both projection and image reconstructed spaces. In the projection space, performance metrics include count rate, energy resolution, uniformity, modulation transfer function (MTF), and noise power spectrum (NPS). In the image space, performance metrics consist of contrast-to-noise ratio (CNR), uniformity, NPS, and iodine quantification accuracy. For both acquisition modes, signal-to-thickness calibration, for gain and beam hardening corrections, is used before image reconstruction. Images are reconstructed via TIGRE CT software using the standard Feldkamp, Davis, and Kress (FDK) filtered back projection algorithm with a Hann filter and reconstructed with a voxel size of 0.081 mm. Material decomposition is performed using a standard image-based method.</p><p><strong>Results: </strong>In the detector space, the application of hardware-based charge summing correction enhances spectral resolution and improves the spatial resolution of MTF at lower energy thresholds but introduces anomalous edge enhancement effects and artifacts in the MTF at high fluence. A negative noise correlation was observed in AC mode-acquired images. As expected, the AC acquisition mode results in a decreased detector count rate. In the image space, NPS results displayed elevated noise in low-energy AC images. However, at high energy, noise was comparable between both modes. Greater uniformity was observed in SP mode-acquired images. The largest disparity was observed in the iodine quantification test, where the AC mode demonstrates a much stronger linear relationship between estimated and true iodine concentrations than the SP mode.</p><p><strong>Conclusion: </strong>The results are specific to the studied system, reconstruction parameters, and irradiation conditions limited to 200 mA and 0.5 mAs. The AC mode generally provides better energy and MTF resolution at low energy thresholds but with increased noise and reduced uniformity. In image space, charge summing correction improved iodine quantification and CNR at high energy thresholds.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 1","pages":"013501"},"PeriodicalIF":1.9,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11759667/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143048266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Toward Continued Growth for the JMI Community.","authors":"","doi":"10.1117/1.JMI.12.1.010101","DOIUrl":"https://doi.org/10.1117/1.JMI.12.1.010101","url":null,"abstract":"<p><p>JMI Editor in Chief Bennett Landman provides an overview of JMI Volume 12 Issue 1 and spotlights key aspects of JMI peer review, with an eye toward continued growth for the JMI community.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 1","pages":"010101"},"PeriodicalIF":1.9,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11817210/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143415687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yassine Oukdach, Anass Garbaz, Zakaria Kerkaou, Mohamed El Ansari, Lahcen Koutti, Nikolaos Papachrysos, Ahmed Fouad El Ouafdi, Thomas de Lange, Cosimo Distante
{"title":"Vision transformer distillation for enhanced gastrointestinal abnormality recognition in wireless capsule endoscopy images.","authors":"Yassine Oukdach, Anass Garbaz, Zakaria Kerkaou, Mohamed El Ansari, Lahcen Koutti, Nikolaos Papachrysos, Ahmed Fouad El Ouafdi, Thomas de Lange, Cosimo Distante","doi":"10.1117/1.JMI.12.1.014505","DOIUrl":"10.1117/1.JMI.12.1.014505","url":null,"abstract":"<p><strong>Purpose: </strong>Wireless capsule endoscopy (WCE) is a non-invasive technology used for diagnosing gastrointestinal abnormalities. A single examination generates <math><mrow><mo>∼</mo> <mn>55,000</mn></mrow> </math> images, making manual review both time-consuming and costly for doctors. Therefore, the development of computer vision-assisted systems is highly desirable to aid in the diagnostic process.</p><p><strong>Approach: </strong>We presents a deep learning approach leveraging knowledge distillation (KD) from a convolutional neural network (CNN) teacher model to a vision transformer (ViT) student model for gastrointestinal abnormality recognition. The CNN teacher model utilizes attention mechanisms and depth-wise separable convolutions to extract features from WCE images, supervising the ViT in learning these representations.</p><p><strong>Results: </strong>The proposed method achieves accuracy of 97% and 96% on the Kvasir and KID datasets, respectively, demonstrating its effectiveness in distinguishing normal from abnormal regions and bleeding from non-bleeding cases. The proposed approach offers computational efficiency and generalization to unseen datasets, outperforming several state-of-the-art methods.</p><p><strong>Conclusions: </strong>We proposed a deep learning approach utilizing CNNs and a ViT with KD to effectively classify gastrointestinal diseases in WCE images. It demonstrates promising performance on public datasets, distinguishing normal from abnormal regions and bleeding from non-bleeding cases while offering optimal computational efficiency compared with existing methods, making it suitable for GI disease applications.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 1","pages":"014505"},"PeriodicalIF":1.9,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11796471/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143366556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Comparing synthetic mammograms based on wide-angle digital breast tomosynthesis with digital mammograms.","authors":"Magnus Dustler, Gustav Hellgren, Pontus Timberg","doi":"10.1117/1.JMI.12.S1.S13011","DOIUrl":"10.1117/1.JMI.12.S1.S13011","url":null,"abstract":"<p><strong>Purpose: </strong>We aim to investigate the characteristics and evaluate the performance of synthetic mammograms (SMs) based on wide-angle digital breast tomosynthesis (DBT) compared with digital mammography (DM).</p><p><strong>Approach: </strong>Fifty cases with both synthetic and digital mammograms were selected from the Malmö Breast Tomosynthesis Screening Trial. They were categorized into five groups consisting of normal cases and recalled cases with false-positive and true-positive findings from DM and DBT only. The DBT system used was a wide-angle (WA) system from Siemens, and the SM images were reconstructed from the DBT images. Visual grading, detection, and recall were evaluated by experienced breast radiologists in both SM and DM images.</p><p><strong>Results: </strong>Some image quality criteria of the SM images were rated as qualitatively inferior to DM images. However, reader-averaged diagnostic accuracy (0.57 versus 0.55), sensitivity (0.46 versus 0.50), and specificity (0.64 versus 0.58) were not significantly different between SM and DM, respectively.</p><p><strong>Conclusions: </strong>Synthetic mammography plays a promising role to complement or even replace DM. The study could not find any indications of substantial differences in the sensitivity or specificity of SM for WA DBT systems compared with DM. However, certain image quality criteria of SM fall slightly short compared with DM images. Next-generation DBT systems could address such limitations through improved reconstruction algorithms and system design, and their performance should be the focus of future research studies.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 Suppl 1","pages":"S13011"},"PeriodicalIF":1.9,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11745418/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143014059","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"OPHash: learning of organ and pathology context-sensitive hashing for medical image retrieval.","authors":"Asim Manna, Rakshith Sathish, Ramanathan Sethuraman, Debdoot Sheet","doi":"10.1117/1.JMI.12.1.017503","DOIUrl":"10.1117/1.JMI.12.1.017503","url":null,"abstract":"<p><strong>Purpose: </strong>Retrieving images of organs and their associated pathologies is essential for evidence-based clinical diagnosis. Deep neural hashing (DNH) has demonstrated the ability to retrieve images fast on large datasets. Conventional pairwise DNH methods can focus on semantic similarity between either organs or pathology of an image pair but not on both simultaneously.</p><p><strong>Approach: </strong>We propose an organ and pathology contextual-supervised hashing approach (OPHash) learned using three types of samples (called bags) to learn accurate hash representation. Because only semantic similarity is inadequate to incorporate with these bags, we introduce relational similarity to generate identical hash codes from most similar image pairs. OPHash is trained by minimizing classification loss, two retrieval losses implemented using Cauchy cross-entropy and maximizing discriminator loss over training samples.</p><p><strong>Results: </strong>Experiments are performed with two radiology datasets derived from the publicly available datasets. OPHash achieves 24% higher mean average precision than the state-of-the-art for top-100 retrieval.</p><p><strong>Conclusion: </strong>OPHash retrieves images with semantic similarity of organs and their associated pathology. It is agnostic to image size as well. This method improves retrieval efficiency across diverse medical imaging datasets, accommodating multiple organs and pathologies. The code is available at https://github.com/asimmanna17/OPHash.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 1","pages":"017503"},"PeriodicalIF":1.9,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11838790/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143469590","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}