Improving realism in abdominal ultrasound simulation combining a segmentation-guided loss and polar coordinates training.

Medical physics Pub Date : 2025-03-30 DOI:10.1002/mp.17801
Santiago Vitale, José Ignacio Orlando, Emmanuel Iarussi, Alejandro Díaz, Ignacio Larrabide
{"title":"Improving realism in abdominal ultrasound simulation combining a segmentation-guided loss and polar coordinates training.","authors":"Santiago Vitale, José Ignacio Orlando, Emmanuel Iarussi, Alejandro Díaz, Ignacio Larrabide","doi":"10.1002/mp.17801","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Ultrasound (US) simulation helps train physicians and medical students in image acquisition and interpretation, enabling safe practice of transducer manipulation and organ identification. Current simulators generate realistic images from reference scans. Although physics-based simulators provide real-time images, they lack sufficient realism, while recent deep learning-based models based on unpaired image-to-image translation improve realism but introduce anatomical inconsistencies.</p><p><strong>Purpose: </strong>We propose a novel framework to reduce hallucinations from generative adversarial networks (GANs) used on physics-based simulations, enhancing anatomical accuracy and realism in abdominal US simulation. Our method aims to produce anatomically consistent images free from artifacts within and outside the field of view (FoV).</p><p><strong>Methods: </strong>We introduce a segmentation-guided loss to enforce anatomical consistency by using a pre-trained Unet model that segments abdominal organs from physics-based simulated scans. Penalizing segmentation discrepancies before and after the translation cycle helps prevent unrealistic artifacts. Additionally, we propose training GANs on images in polar coordinates to limit the field of view to non-blank regions. We evaluated our approach on unpaired datasets comprising 617 real abdominal US images from a SonoSite-M turbo v1.3 scanner and 971 artificial scans from a ray-casting simulator. Data was partitioned at the patient level into training (70%), validation (10%), and testing (20%). Performance was quantitatively assessed with Frechet and Kernel Inception Distances (FID and KID), and organ-specific <math> <semantics><msup><mi>χ</mi> <mn>2</mn></msup> <annotation>$\\chi ^2$</annotation></semantics> </math> histogram distances, reporting 95% confidence intervals. We compared our model against generative methods such as CUT, UVCGANv2, and UNSB, performing statistical analyses using Wilcoxon tests (FID and KID with Bonferroni-corrected <math> <semantics><mrow><mi>α</mi> <mo>=</mo> <mn>0.01</mn></mrow> <annotation>$\\alpha = 0.01$</annotation></semantics> </math> , <math> <semantics><msup><mi>χ</mi> <mn>2</mn></msup> <annotation>$\\chi ^2$</annotation></semantics> </math> with <math> <semantics><mrow><mi>α</mi> <mo>=</mo> <mn>0.008</mn></mrow> <annotation>$\\alpha =0.008$</annotation></semantics> </math> ). A perceptual realism study involving expert radiologists was also conducted.</p><p><strong>Results: </strong>Our method significantly reduced FID and KID by 66% and 89%, respectively, compared to CycleGAN, and by 34% and 59% compared to the leading alternative UVCGANv2 ( <math> <semantics><mrow><mi>p</mi> <mo>≪</mo> <mn>0.01</mn></mrow> <annotation>$p \\ll 0.01$</annotation></semantics> </math> ). No significant differences ( <math> <semantics><mrow><mi>p</mi> <mo>></mo> <mn>0.008</mn></mrow> <annotation>$p>0.008$</annotation></semantics> </math> ) in echogenicity distributions were found between real and simulated images within liver and gallbladder regions. The user study indicated our simulated scans fooled radiologists in 36.2% of cases, outperforming other methods.</p><p><strong>Conclusions: </strong>Our segmentation-guided, polar-coordinates-trained CycleGAN framework significantly reduces hallucinations, ensuring anatomical consistency, and realism in simulated abdominal US images, surpassing existing methods.</p>","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2025-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical physics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1002/mp.17801","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Ultrasound (US) simulation helps train physicians and medical students in image acquisition and interpretation, enabling safe practice of transducer manipulation and organ identification. Current simulators generate realistic images from reference scans. Although physics-based simulators provide real-time images, they lack sufficient realism, while recent deep learning-based models based on unpaired image-to-image translation improve realism but introduce anatomical inconsistencies.

Purpose: We propose a novel framework to reduce hallucinations from generative adversarial networks (GANs) used on physics-based simulations, enhancing anatomical accuracy and realism in abdominal US simulation. Our method aims to produce anatomically consistent images free from artifacts within and outside the field of view (FoV).

Methods: We introduce a segmentation-guided loss to enforce anatomical consistency by using a pre-trained Unet model that segments abdominal organs from physics-based simulated scans. Penalizing segmentation discrepancies before and after the translation cycle helps prevent unrealistic artifacts. Additionally, we propose training GANs on images in polar coordinates to limit the field of view to non-blank regions. We evaluated our approach on unpaired datasets comprising 617 real abdominal US images from a SonoSite-M turbo v1.3 scanner and 971 artificial scans from a ray-casting simulator. Data was partitioned at the patient level into training (70%), validation (10%), and testing (20%). Performance was quantitatively assessed with Frechet and Kernel Inception Distances (FID and KID), and organ-specific χ 2 $\chi ^2$ histogram distances, reporting 95% confidence intervals. We compared our model against generative methods such as CUT, UVCGANv2, and UNSB, performing statistical analyses using Wilcoxon tests (FID and KID with Bonferroni-corrected α = 0.01 $\alpha = 0.01$ , χ 2 $\chi ^2$ with α = 0.008 $\alpha =0.008$ ). A perceptual realism study involving expert radiologists was also conducted.

Results: Our method significantly reduced FID and KID by 66% and 89%, respectively, compared to CycleGAN, and by 34% and 59% compared to the leading alternative UVCGANv2 ( p 0.01 $p \ll 0.01$ ). No significant differences ( p > 0.008 $p>0.008$ ) in echogenicity distributions were found between real and simulated images within liver and gallbladder regions. The user study indicated our simulated scans fooled radiologists in 36.2% of cases, outperforming other methods.

Conclusions: Our segmentation-guided, polar-coordinates-trained CycleGAN framework significantly reduces hallucinations, ensuring anatomical consistency, and realism in simulated abdominal US images, surpassing existing methods.

结合分割引导损失和极坐标训练提高腹部超声模拟的逼真度。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信