Hongjiang Lu, Miao Liu, Kun Yu, Yuan Fang, Jing Zhao, Yang Shi
{"title":"基于深度学习的全自动椎体分割和标记工作流程。","authors":"Hongjiang Lu, Miao Liu, Kun Yu, Yuan Fang, Jing Zhao, Yang Shi","doi":"10.12968/hmed.2025.0443","DOIUrl":null,"url":null,"abstract":"<p><p><b>Aims/Background</b> Spinal disorders, such as herniated discs and scoliosis, are highly prevalent conditions with rising incidence in the aging global population. Accurate analysis of spinal anatomical structures is a critical prerequisite for achieving high-precision positioning with surgical navigation robots. However, traditional manual segmentation methods are limited by issues such as low efficiency and poor consistency. This work aims to develop a fully automated deep learning-based vertebral segmentation and labeling workflow to provide efficient and accurate preoperative analysis support for spine surgery navigation robots. <b>Methods</b> In the localization stage, the You Only Look Once version 7 (YOLOv7) network was utilized to predict the bounding boxes of individual vertebrae on computed tomography (CT) sagittal slices, transforming the 3D localization problem into a 2D one. Subsequently, the Density-Based Spatial Clustering of Applications with Noise (DBSCAN) clustering algorithm was employed to aggregate the 2D detection results into 3D vertebral centers. This approach significantly reduces inference time and enhances localization accuracy. In the segmentation stage, a 3D U-Net model integrated with an attention mechanism was trained using the region of interest (ROI) based on the vertebral center as input, effectively extracting the 3D structural features of vertebrae to achieve precise segmentation. In the labeling stage, a vertebra labeling network was trained by combining deep learning architectures-ResNet and Transformer, which are capable of extracting rich intervertebral features, to obtain the final labeling results through post-processing based on positional logic analysis. To verify the effectiveness of this workflow, experiments were conducted on a dataset comprising 106 spinal CT datasets sourced from various devices, covering a wide range of clinical scenarios. <b>Results</b> The results demonstrate that the method performed excellently in the three key tasks of localization, segmentation, and labeling, with a Mean Localization Error (MLE) of 1.42 mm. The segmentation accuracy metrics included a Dice Similarity Coefficient (DSC) of 0.968 ± 0.014, Intersection over Union (IoU) of 0.879 ± 0.018, Pixel Accuracy (PA) of 0.988 ± 0.005, mean symmetric distance (MSD) of 1.09 ± 0.19 mm, and Hausdorff Distance (HD) of 5.42 ± 2.05 mm. The degree of classification accuracy reached up to 94.36%. <b>Conclusion</b> These quantitative assessments and visualizations confirm the effectiveness of our method (vertebra localization, vertebra segmentation and vertebra labeling), indicating its potential for deployment in spinal surgery navigation robots to provide accurate and efficient preoperative analysis and navigation support for spinal surgeries.</p>","PeriodicalId":9256,"journal":{"name":"British journal of hospital medicine","volume":"86 9","pages":"1-22"},"PeriodicalIF":1.8000,"publicationDate":"2025-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Deep Learning-Based Fully Automated Vertebra Segmentation and Labeling Workflow.\",\"authors\":\"Hongjiang Lu, Miao Liu, Kun Yu, Yuan Fang, Jing Zhao, Yang Shi\",\"doi\":\"10.12968/hmed.2025.0443\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p><b>Aims/Background</b> Spinal disorders, such as herniated discs and scoliosis, are highly prevalent conditions with rising incidence in the aging global population. Accurate analysis of spinal anatomical structures is a critical prerequisite for achieving high-precision positioning with surgical navigation robots. However, traditional manual segmentation methods are limited by issues such as low efficiency and poor consistency. This work aims to develop a fully automated deep learning-based vertebral segmentation and labeling workflow to provide efficient and accurate preoperative analysis support for spine surgery navigation robots. <b>Methods</b> In the localization stage, the You Only Look Once version 7 (YOLOv7) network was utilized to predict the bounding boxes of individual vertebrae on computed tomography (CT) sagittal slices, transforming the 3D localization problem into a 2D one. Subsequently, the Density-Based Spatial Clustering of Applications with Noise (DBSCAN) clustering algorithm was employed to aggregate the 2D detection results into 3D vertebral centers. This approach significantly reduces inference time and enhances localization accuracy. In the segmentation stage, a 3D U-Net model integrated with an attention mechanism was trained using the region of interest (ROI) based on the vertebral center as input, effectively extracting the 3D structural features of vertebrae to achieve precise segmentation. In the labeling stage, a vertebra labeling network was trained by combining deep learning architectures-ResNet and Transformer, which are capable of extracting rich intervertebral features, to obtain the final labeling results through post-processing based on positional logic analysis. To verify the effectiveness of this workflow, experiments were conducted on a dataset comprising 106 spinal CT datasets sourced from various devices, covering a wide range of clinical scenarios. <b>Results</b> The results demonstrate that the method performed excellently in the three key tasks of localization, segmentation, and labeling, with a Mean Localization Error (MLE) of 1.42 mm. The segmentation accuracy metrics included a Dice Similarity Coefficient (DSC) of 0.968 ± 0.014, Intersection over Union (IoU) of 0.879 ± 0.018, Pixel Accuracy (PA) of 0.988 ± 0.005, mean symmetric distance (MSD) of 1.09 ± 0.19 mm, and Hausdorff Distance (HD) of 5.42 ± 2.05 mm. The degree of classification accuracy reached up to 94.36%. <b>Conclusion</b> These quantitative assessments and visualizations confirm the effectiveness of our method (vertebra localization, vertebra segmentation and vertebra labeling), indicating its potential for deployment in spinal surgery navigation robots to provide accurate and efficient preoperative analysis and navigation support for spinal surgeries.</p>\",\"PeriodicalId\":9256,\"journal\":{\"name\":\"British journal of hospital medicine\",\"volume\":\"86 9\",\"pages\":\"1-22\"},\"PeriodicalIF\":1.8000,\"publicationDate\":\"2025-09-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"British journal of hospital medicine\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.12968/hmed.2025.0443\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/9/24 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"Q3\",\"JCRName\":\"MEDICINE, GENERAL & INTERNAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"British journal of hospital medicine","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.12968/hmed.2025.0443","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/9/24 0:00:00","PubModel":"Epub","JCR":"Q3","JCRName":"MEDICINE, GENERAL & INTERNAL","Score":null,"Total":0}
引用次数: 0
摘要
目的/背景脊柱疾病,如椎间盘突出和脊柱侧凸,是全球老龄化人口中发病率不断上升的非常普遍的疾病。准确分析脊柱解剖结构是实现手术导航机器人高精度定位的关键前提。然而,传统的人工分割方法存在效率低、一致性差等问题。本工作旨在开发一种全自动的基于深度学习的椎体分割和标记工作流程,为脊柱手术导航机器人提供高效准确的术前分析支持。方法在定位阶段,利用You Only Look Once version 7 (YOLOv7)网络预测单个椎体在CT矢状面切片上的边界框,将三维定位问题转化为二维定位问题。随后,采用基于密度的带噪声应用空间聚类(DBSCAN)聚类算法,将二维检测结果聚合为三维椎体中心。该方法显著缩短了推理时间,提高了定位精度。在分割阶段,以基于椎体中心的感兴趣区域(region of interest, ROI)为输入,训练集成注意机制的三维U-Net模型,有效提取椎体的三维结构特征,实现精确分割。在标记阶段,结合能够提取丰富椎间特征的深度学习架构resnet和Transformer,训练一个椎体标记网络,通过基于位置逻辑分析的后处理,得到最终的标记结果。为了验证该工作流程的有效性,我们在一个包含106个脊柱CT数据集的数据集上进行了实验,这些数据集来自不同的设备,涵盖了广泛的临床场景。结果该方法在定位、分割和标记三个关键任务上表现优异,平均定位误差(MLE)为1.42 mm。分割精度指标为:Dice Similarity Coefficient (DSC) 0.968±0.014,Intersection over Union (IoU) 0.879±0.018,Pixel accuracy (PA) 0.988±0.005,平均对称距离(MSD) 1.09±0.19 mm, Hausdorff distance (HD) 5.42±2.05 mm。分类准确率达到94.36%。结论这些定量评估和可视化结果证实了我们的方法(椎体定位、椎体分割和椎体标记)的有效性,表明了该方法在脊柱手术导航机器人中的应用潜力,为脊柱手术提供准确、高效的术前分析和导航支持。
A Deep Learning-Based Fully Automated Vertebra Segmentation and Labeling Workflow.
Aims/Background Spinal disorders, such as herniated discs and scoliosis, are highly prevalent conditions with rising incidence in the aging global population. Accurate analysis of spinal anatomical structures is a critical prerequisite for achieving high-precision positioning with surgical navigation robots. However, traditional manual segmentation methods are limited by issues such as low efficiency and poor consistency. This work aims to develop a fully automated deep learning-based vertebral segmentation and labeling workflow to provide efficient and accurate preoperative analysis support for spine surgery navigation robots. Methods In the localization stage, the You Only Look Once version 7 (YOLOv7) network was utilized to predict the bounding boxes of individual vertebrae on computed tomography (CT) sagittal slices, transforming the 3D localization problem into a 2D one. Subsequently, the Density-Based Spatial Clustering of Applications with Noise (DBSCAN) clustering algorithm was employed to aggregate the 2D detection results into 3D vertebral centers. This approach significantly reduces inference time and enhances localization accuracy. In the segmentation stage, a 3D U-Net model integrated with an attention mechanism was trained using the region of interest (ROI) based on the vertebral center as input, effectively extracting the 3D structural features of vertebrae to achieve precise segmentation. In the labeling stage, a vertebra labeling network was trained by combining deep learning architectures-ResNet and Transformer, which are capable of extracting rich intervertebral features, to obtain the final labeling results through post-processing based on positional logic analysis. To verify the effectiveness of this workflow, experiments were conducted on a dataset comprising 106 spinal CT datasets sourced from various devices, covering a wide range of clinical scenarios. Results The results demonstrate that the method performed excellently in the three key tasks of localization, segmentation, and labeling, with a Mean Localization Error (MLE) of 1.42 mm. The segmentation accuracy metrics included a Dice Similarity Coefficient (DSC) of 0.968 ± 0.014, Intersection over Union (IoU) of 0.879 ± 0.018, Pixel Accuracy (PA) of 0.988 ± 0.005, mean symmetric distance (MSD) of 1.09 ± 0.19 mm, and Hausdorff Distance (HD) of 5.42 ± 2.05 mm. The degree of classification accuracy reached up to 94.36%. Conclusion These quantitative assessments and visualizations confirm the effectiveness of our method (vertebra localization, vertebra segmentation and vertebra labeling), indicating its potential for deployment in spinal surgery navigation robots to provide accurate and efficient preoperative analysis and navigation support for spinal surgeries.
期刊介绍:
British Journal of Hospital Medicine was established in 1966, and is still true to its origins: a monthly, peer-reviewed, multidisciplinary review journal for hospital doctors and doctors in training.
The journal publishes an authoritative mix of clinical reviews, education and training updates, quality improvement projects and case reports, and book reviews from recognized leaders in the profession. The Core Training for Doctors section provides clinical information in an easily accessible format for doctors in training.
British Journal of Hospital Medicine is an invaluable resource for hospital doctors at all stages of their career.
The journal is indexed on Medline, CINAHL, the Sociedad Iberoamericana de Información Científica and Scopus.