{"title":"GSV-Pose:基于几何相似度投票的姿态估计","authors":"Xi Zhao, Yuekun Zhang, Jinji Wu","doi":"10.1007/s10489-025-06853-z","DOIUrl":null,"url":null,"abstract":"<div><p>Object pose estimation is a fundamental problem in 3D computer vision and has gained significant attention with the rapid advancements in autonomous driving, robotics, and augmented reality. Traditional voting-based approaches often suffer from reduced accuracy when dealing with partially observed objects. To overcome this limitation, our method incorporates a superpoint matching network to compute local geometric similarities, which effectively guides the voting process and enhances pose estimation robustness. Experimental results demonstrate that our approach achieves comparable performance to the current state-of-the-art (SOTA) method, GPV-Pose, under standard conditions. More importantly, in robustness tests with incomplete objects, our method significantly surpasses GPV-Pose. For instance, under a 20% incompleteness ratio, the accuracy of GPV-Pose drops by 61.6% under the <span>\\(5^{\\circ }2\\,\\text {cm}\\)</span> criterion, whereas our method experiences only a 21.8% reduction.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 15","pages":""},"PeriodicalIF":3.5000,"publicationDate":"2025-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"GSV-Pose: Pose estimation based on geometric similarity voting\",\"authors\":\"Xi Zhao, Yuekun Zhang, Jinji Wu\",\"doi\":\"10.1007/s10489-025-06853-z\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Object pose estimation is a fundamental problem in 3D computer vision and has gained significant attention with the rapid advancements in autonomous driving, robotics, and augmented reality. Traditional voting-based approaches often suffer from reduced accuracy when dealing with partially observed objects. To overcome this limitation, our method incorporates a superpoint matching network to compute local geometric similarities, which effectively guides the voting process and enhances pose estimation robustness. Experimental results demonstrate that our approach achieves comparable performance to the current state-of-the-art (SOTA) method, GPV-Pose, under standard conditions. More importantly, in robustness tests with incomplete objects, our method significantly surpasses GPV-Pose. For instance, under a 20% incompleteness ratio, the accuracy of GPV-Pose drops by 61.6% under the <span>\\\\(5^{\\\\circ }2\\\\,\\\\text {cm}\\\\)</span> criterion, whereas our method experiences only a 21.8% reduction.</p></div>\",\"PeriodicalId\":8041,\"journal\":{\"name\":\"Applied Intelligence\",\"volume\":\"55 15\",\"pages\":\"\"},\"PeriodicalIF\":3.5000,\"publicationDate\":\"2025-10-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://link.springer.com/article/10.1007/s10489-025-06853-z\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Intelligence","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10489-025-06853-z","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
摘要
物体姿态估计是三维计算机视觉中的一个基本问题,随着自动驾驶、机器人技术和增强现实技术的快速发展,该问题受到了广泛关注。传统的基于投票的方法在处理部分观察到的对象时往往存在准确性降低的问题。为了克服这一限制,我们的方法引入了一个多点匹配网络来计算局部几何相似度,有效地指导了投票过程,增强了姿态估计的鲁棒性。实验结果表明,在标准条件下,我们的方法达到了与当前最先进的(SOTA)方法GPV-Pose相当的性能。更重要的是,在不完整对象的鲁棒性测试中,我们的方法明显优于GPV-Pose。例如,低于20% incompleteness ratio, the accuracy of GPV-Pose drops by 61.6% under the \(5^{\circ }2\,\text {cm}\) criterion, whereas our method experiences only a 21.8% reduction.
GSV-Pose: Pose estimation based on geometric similarity voting
Object pose estimation is a fundamental problem in 3D computer vision and has gained significant attention with the rapid advancements in autonomous driving, robotics, and augmented reality. Traditional voting-based approaches often suffer from reduced accuracy when dealing with partially observed objects. To overcome this limitation, our method incorporates a superpoint matching network to compute local geometric similarities, which effectively guides the voting process and enhances pose estimation robustness. Experimental results demonstrate that our approach achieves comparable performance to the current state-of-the-art (SOTA) method, GPV-Pose, under standard conditions. More importantly, in robustness tests with incomplete objects, our method significantly surpasses GPV-Pose. For instance, under a 20% incompleteness ratio, the accuracy of GPV-Pose drops by 61.6% under the \(5^{\circ }2\,\text {cm}\) criterion, whereas our method experiences only a 21.8% reduction.
期刊介绍:
With a focus on research in artificial intelligence and neural networks, this journal addresses issues involving solutions of real-life manufacturing, defense, management, government and industrial problems which are too complex to be solved through conventional approaches and require the simulation of intelligent thought processes, heuristics, applications of knowledge, and distributed and parallel processing. The integration of these multiple approaches in solving complex problems is of particular importance.
The journal presents new and original research and technological developments, addressing real and complex issues applicable to difficult problems. It provides a medium for exchanging scientific research and technological achievements accomplished by the international community.