{"title":"Monocular suture needle pose detection using synthetic data augmented convolutional neural network.","authors":"Yifan Wang, Saul Alexis Heredia Perez, Kanako Harada","doi":"10.1007/s11548-025-03467-1","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>Robotic microsurgery enhances the dexterity and stability of the surgeon to perform precise and delicate surgical procedures at microscopic level. Accurate needle pose estimation is critical for robotic micro-suturing, enabling optimized insertion trajectories and facilitating autonomous control. However, accurately estimating the pose of a needle during manipulation, particularly under monocular vision, remains a challenge. This study proposes a convolutional neural network-based method to estimate the pose of a suture needle from monocular images.</p><p><strong>Methods: </strong>The 3D pose of the needle is estimated using keypoints information from 2D images. A convolutional neural network was trained to estimate the positions of keypoints on the needle, specifically the tip, middle and end point. A hybrid dataset comprising images from both real-world and synthetic simulated environments was developed to train the model. Subsequently, an algorithm was designed to estimate the 3D positions of these keypoints. The 2D keypoint detection and 3D orientation estimation were evaluated by translation and orientation error metrics, respectively.</p><p><strong>Results: </strong>Experiments conducted on synthetic data showed that the average translation error of tip point, middle point and end point being 0.107 mm, 0.118 mm and 0.098 mm, and the average orientation angular error was 12.75 <math><mmultiscripts><mrow></mrow> <mrow></mrow> <mo>∘</mo></mmultiscripts> </math> for normal vector and 15.55 <math><mmultiscripts><mrow></mrow> <mrow></mrow> <mo>∘</mo></mmultiscripts> </math> for direction vector. When evaluated on real data, the method demonstrated 2D translation errors averaging 0.047 mm, 0.052 mm and 0.049 mm for the respective keypoints, with 93.85% of detected keypoints having errors below 4 pixels.</p><p><strong>Conclusions: </strong>This study presents a CNN-based method, augmented with synthetic images, to estimate the pose of a suture needle in monocular vision. Experimental results indicate that the method effectively estimates the 2D positions and 3D orientations of the suture needle in synthetic images. The model also shows reasonable performance with real data, highlighting its promise for real-time application in robotic microsurgery.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2019-2030"},"PeriodicalIF":2.3000,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12518471/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Computer Assisted Radiology and Surgery","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1007/s11548-025-03467-1","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/6/24 0:00:00","PubModel":"Epub","JCR":"Q3","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Purpose: Robotic microsurgery enhances the dexterity and stability of the surgeon to perform precise and delicate surgical procedures at microscopic level. Accurate needle pose estimation is critical for robotic micro-suturing, enabling optimized insertion trajectories and facilitating autonomous control. However, accurately estimating the pose of a needle during manipulation, particularly under monocular vision, remains a challenge. This study proposes a convolutional neural network-based method to estimate the pose of a suture needle from monocular images.
Methods: The 3D pose of the needle is estimated using keypoints information from 2D images. A convolutional neural network was trained to estimate the positions of keypoints on the needle, specifically the tip, middle and end point. A hybrid dataset comprising images from both real-world and synthetic simulated environments was developed to train the model. Subsequently, an algorithm was designed to estimate the 3D positions of these keypoints. The 2D keypoint detection and 3D orientation estimation were evaluated by translation and orientation error metrics, respectively.
Results: Experiments conducted on synthetic data showed that the average translation error of tip point, middle point and end point being 0.107 mm, 0.118 mm and 0.098 mm, and the average orientation angular error was 12.75 for normal vector and 15.55 for direction vector. When evaluated on real data, the method demonstrated 2D translation errors averaging 0.047 mm, 0.052 mm and 0.049 mm for the respective keypoints, with 93.85% of detected keypoints having errors below 4 pixels.
Conclusions: This study presents a CNN-based method, augmented with synthetic images, to estimate the pose of a suture needle in monocular vision. Experimental results indicate that the method effectively estimates the 2D positions and 3D orientations of the suture needle in synthetic images. The model also shows reasonable performance with real data, highlighting its promise for real-time application in robotic microsurgery.
期刊介绍:
The International Journal for Computer Assisted Radiology and Surgery (IJCARS) is a peer-reviewed journal that provides a platform for closing the gap between medical and technical disciplines, and encourages interdisciplinary research and development activities in an international environment.