Domain adaptation strategies for 3D reconstruction of the lumbar spine using real fluoroscopy data

IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Sascha Jecklin , Youyang Shen , Amandine Gout , Daniel Suter , Lilian Calvet , Lukas Zingg , Jennifer Straub , Nicola Alessandro Cavalcanti , Mazda Farshad , Philipp Fürnstahl , Hooman Esfandiari
{"title":"Domain adaptation strategies for 3D reconstruction of the lumbar spine using real fluoroscopy data","authors":"Sascha Jecklin ,&nbsp;Youyang Shen ,&nbsp;Amandine Gout ,&nbsp;Daniel Suter ,&nbsp;Lilian Calvet ,&nbsp;Lukas Zingg ,&nbsp;Jennifer Straub ,&nbsp;Nicola Alessandro Cavalcanti ,&nbsp;Mazda Farshad ,&nbsp;Philipp Fürnstahl ,&nbsp;Hooman Esfandiari","doi":"10.1016/j.media.2024.103322","DOIUrl":null,"url":null,"abstract":"<div><p>In this study, we address critical barriers hindering the widespread adoption of surgical navigation in orthopedic surgeries due to limitations such as time constraints, cost implications, radiation concerns, and integration within the surgical workflow. Recently, our work X23D showed an approach for generating 3D anatomical models of the spine from only a few intraoperative fluoroscopic images. This approach negates the need for conventional registration-based surgical navigation by creating a direct intraoperative 3D reconstruction of the anatomy. Despite these strides, the practical application of X23D has been limited by a significant domain gap between synthetic training data and real intraoperative images.</p><p>In response, we devised a novel data collection protocol to assemble a paired dataset consisting of synthetic and real fluoroscopic images captured from identical perspectives. Leveraging this unique dataset, we refined our deep learning model through transfer learning, effectively bridging the domain gap between synthetic and real X-ray data. We introduce an innovative approach combining style transfer with the curated paired dataset. This method transforms real X-ray images into the synthetic domain, enabling the <em>in-silico</em>-trained X23D model to achieve high accuracy in real-world settings.</p><p>Our results demonstrated that the refined model can rapidly generate accurate 3D reconstructions of the entire lumbar spine from as few as three intraoperative fluoroscopic shots. The enhanced model reached a sufficient accuracy, achieving an 84% F1 score, equating to the benchmark set solely by synthetic data in previous research. Moreover, with an impressive computational time of just 81.1 ms, our approach offers real-time capabilities, vital for successful integration into active surgical procedures.</p><p>By investigating optimal imaging setups and view angle dependencies, we have further validated the practicality and reliability of our system in a clinical environment. Our research represents a promising advancement in intraoperative 3D reconstruction. This innovation has the potential to enhance intraoperative surgical planning, navigation, and surgical robotics.</p></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"98 ","pages":"Article 103322"},"PeriodicalIF":10.7000,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1361841524002470/pdfft?md5=c8d17bbbaa45287c29e84ed636f09188&pid=1-s2.0-S1361841524002470-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical image analysis","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1361841524002470","RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

In this study, we address critical barriers hindering the widespread adoption of surgical navigation in orthopedic surgeries due to limitations such as time constraints, cost implications, radiation concerns, and integration within the surgical workflow. Recently, our work X23D showed an approach for generating 3D anatomical models of the spine from only a few intraoperative fluoroscopic images. This approach negates the need for conventional registration-based surgical navigation by creating a direct intraoperative 3D reconstruction of the anatomy. Despite these strides, the practical application of X23D has been limited by a significant domain gap between synthetic training data and real intraoperative images.

In response, we devised a novel data collection protocol to assemble a paired dataset consisting of synthetic and real fluoroscopic images captured from identical perspectives. Leveraging this unique dataset, we refined our deep learning model through transfer learning, effectively bridging the domain gap between synthetic and real X-ray data. We introduce an innovative approach combining style transfer with the curated paired dataset. This method transforms real X-ray images into the synthetic domain, enabling the in-silico-trained X23D model to achieve high accuracy in real-world settings.

Our results demonstrated that the refined model can rapidly generate accurate 3D reconstructions of the entire lumbar spine from as few as three intraoperative fluoroscopic shots. The enhanced model reached a sufficient accuracy, achieving an 84% F1 score, equating to the benchmark set solely by synthetic data in previous research. Moreover, with an impressive computational time of just 81.1 ms, our approach offers real-time capabilities, vital for successful integration into active surgical procedures.

By investigating optimal imaging setups and view angle dependencies, we have further validated the practicality and reliability of our system in a clinical environment. Our research represents a promising advancement in intraoperative 3D reconstruction. This innovation has the potential to enhance intraoperative surgical planning, navigation, and surgical robotics.

利用真实透视数据进行腰椎三维重建的领域适应策略
在本研究中,我们探讨了骨科手术中广泛采用手术导航的关键障碍,这些障碍是由于时间限制、成本影响、辐射问题以及手术流程整合等限制因素造成的。最近,我们的工作 X23D 展示了一种仅通过几张术中透视图像就能生成脊柱三维解剖模型的方法。这种方法通过在术中直接创建解剖结构的三维重建,无需传统的基于配准的手术导航。尽管取得了这些进展,X23D 的实际应用仍受到合成训练数据与真实术中图像之间巨大的领域差距的限制。为此,我们设计了一种新颖的数据收集方案,将从相同角度拍摄的合成透视图像和真实透视图像组成一个配对数据集。利用这个独特的数据集,我们通过迁移学习完善了深度学习模型,有效弥合了合成和真实 X 射线数据之间的领域差距。我们引入了一种创新方法,将风格转移与策划的配对数据集相结合。我们的研究结果表明,改进后的模型可以从少至三个术中透视镜头快速生成整个腰椎的精确三维重建。改进后的模型达到了足够的精确度,F1得分率为84%,与之前研究中仅通过合成数据设定的基准相当。通过研究最佳成像设置和视角相关性,我们进一步验证了我们的系统在临床环境中的实用性和可靠性。我们的研究代表着术中三维重建技术的巨大进步。这项创新有望增强术中手术规划、导航和手术机器人技术。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Medical image analysis
Medical image analysis 工程技术-工程:生物医学
CiteScore
22.10
自引率
6.40%
发文量
309
审稿时长
6.6 months
期刊介绍: Medical Image Analysis serves as a platform for sharing new research findings in the realm of medical and biological image analysis, with a focus on applications of computer vision, virtual reality, and robotics to biomedical imaging challenges. The journal prioritizes the publication of high-quality, original papers contributing to the fundamental science of processing, analyzing, and utilizing medical and biological images. It welcomes approaches utilizing biomedical image datasets across all spatial scales, from molecular/cellular imaging to tissue/organ imaging.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信