Weili Fang , Guanghui Geng , Gan Zhang , Peter E.D. Love
{"title":"自主发射龙门:用于预制混凝土梁实时姿态估计的改进单目视觉方法","authors":"Weili Fang , Guanghui Geng , Gan Zhang , Peter E.D. Love","doi":"10.1016/j.autcon.2025.106534","DOIUrl":null,"url":null,"abstract":"<div><div>The absence of accurate and real-time 6-DoF pose data for precast concrete girders renders launching gantry operations predominantly manual, thereby impeding further automation. Such limitations pose a critical question: <em>How can we accurately and robustly estimate the pose of precast concrete girders in real-time during launching gantry operations?</em> To address that question, our paper proposes a monocular vision-based approach to estimate the 6-DoF pose of the precast concrete girder in launching gantry operations. The approach detects the ChArUco board regions using the YOLOv11n model, applies GAN-based image deblurring. The 6-DoF pose is then estimated using a Perspective-n-Point solver and transformed to the gantry coordinate system. Field tests demonstrate robust performance, achieving a mean reprojection error of 0.113 pixels and a processing latency of 60 ms per frame. The results validate the approach's robustness and real-time performance, highlighting monocular vision as a cost-effective alternative to LiDAR–IMU fusion for large-scale automation in construction.</div></div>","PeriodicalId":8660,"journal":{"name":"Automation in Construction","volume":"180 ","pages":"Article 106534"},"PeriodicalIF":11.5000,"publicationDate":"2025-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Autonomous launching gantry: Improved monocular vision approach for real-time pose estimation of precast concrete girders\",\"authors\":\"Weili Fang , Guanghui Geng , Gan Zhang , Peter E.D. Love\",\"doi\":\"10.1016/j.autcon.2025.106534\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The absence of accurate and real-time 6-DoF pose data for precast concrete girders renders launching gantry operations predominantly manual, thereby impeding further automation. Such limitations pose a critical question: <em>How can we accurately and robustly estimate the pose of precast concrete girders in real-time during launching gantry operations?</em> To address that question, our paper proposes a monocular vision-based approach to estimate the 6-DoF pose of the precast concrete girder in launching gantry operations. The approach detects the ChArUco board regions using the YOLOv11n model, applies GAN-based image deblurring. The 6-DoF pose is then estimated using a Perspective-n-Point solver and transformed to the gantry coordinate system. Field tests demonstrate robust performance, achieving a mean reprojection error of 0.113 pixels and a processing latency of 60 ms per frame. The results validate the approach's robustness and real-time performance, highlighting monocular vision as a cost-effective alternative to LiDAR–IMU fusion for large-scale automation in construction.</div></div>\",\"PeriodicalId\":8660,\"journal\":{\"name\":\"Automation in Construction\",\"volume\":\"180 \",\"pages\":\"Article 106534\"},\"PeriodicalIF\":11.5000,\"publicationDate\":\"2025-09-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Automation in Construction\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0926580525005746\",\"RegionNum\":1,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"CONSTRUCTION & BUILDING TECHNOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Automation in Construction","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0926580525005746","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"CONSTRUCTION & BUILDING TECHNOLOGY","Score":null,"Total":0}
Autonomous launching gantry: Improved monocular vision approach for real-time pose estimation of precast concrete girders
The absence of accurate and real-time 6-DoF pose data for precast concrete girders renders launching gantry operations predominantly manual, thereby impeding further automation. Such limitations pose a critical question: How can we accurately and robustly estimate the pose of precast concrete girders in real-time during launching gantry operations? To address that question, our paper proposes a monocular vision-based approach to estimate the 6-DoF pose of the precast concrete girder in launching gantry operations. The approach detects the ChArUco board regions using the YOLOv11n model, applies GAN-based image deblurring. The 6-DoF pose is then estimated using a Perspective-n-Point solver and transformed to the gantry coordinate system. Field tests demonstrate robust performance, achieving a mean reprojection error of 0.113 pixels and a processing latency of 60 ms per frame. The results validate the approach's robustness and real-time performance, highlighting monocular vision as a cost-effective alternative to LiDAR–IMU fusion for large-scale automation in construction.
期刊介绍:
Automation in Construction is an international journal that focuses on publishing original research papers related to the use of Information Technologies in various aspects of the construction industry. The journal covers topics such as design, engineering, construction technologies, and the maintenance and management of constructed facilities.
The scope of Automation in Construction is extensive and covers all stages of the construction life cycle. This includes initial planning and design, construction of the facility, operation and maintenance, as well as the eventual dismantling and recycling of buildings and engineering structures.