{"title":"Anchor-Based 6D Object Pose Estimation","authors":"Zehao Liu, Hao Wang, Fuchang Liu","doi":"10.1109/ICVR51878.2021.9483838","DOIUrl":null,"url":null,"abstract":"Estimating the 6D pose of known objects is an important task for many robotic and AR applications. The most recent trend in 6D pose estimation has been to take advantage of deep networks to either directly regress the pose from the image or to first predict the 2D locations of 3D keypoints and then recover the pose according to the 2D-3D coordinates relationship of keypoints using a PnP algorithm. However, most of these methods are vulnerable to occlusions. In this paper, we present a simple but efficient method to estimate 6D pose of objects using anchor-based corner detection, based on two-stage detection backbone (i.e. Faster R-CNN Ren et al. (2015)). Instead of directly predicting two-dimensional coordinates of the projected 3D bounding box (i.e. corners), we regress relative coordinates of top corners based on the two-dimensional anchor and diagonals of corresponding faces. Bottom corners are further robustly inferred using geometrical constraints of face diagonals. Experiments show that our method achieves significant improvement on LineMod S. Hinterstoisser et al. (2012) and Occlusion Brachmann et al. (2014a) dataset, outperforming most existing 6D pose estimation methods without using refinement.","PeriodicalId":266506,"journal":{"name":"2021 IEEE 7th International Conference on Virtual Reality (ICVR)","volume":"43 11","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 7th International Conference on Virtual Reality (ICVR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICVR51878.2021.9483838","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Estimating the 6D pose of known objects is an important task for many robotic and AR applications. The most recent trend in 6D pose estimation has been to take advantage of deep networks to either directly regress the pose from the image or to first predict the 2D locations of 3D keypoints and then recover the pose according to the 2D-3D coordinates relationship of keypoints using a PnP algorithm. However, most of these methods are vulnerable to occlusions. In this paper, we present a simple but efficient method to estimate 6D pose of objects using anchor-based corner detection, based on two-stage detection backbone (i.e. Faster R-CNN Ren et al. (2015)). Instead of directly predicting two-dimensional coordinates of the projected 3D bounding box (i.e. corners), we regress relative coordinates of top corners based on the two-dimensional anchor and diagonals of corresponding faces. Bottom corners are further robustly inferred using geometrical constraints of face diagonals. Experiments show that our method achieves significant improvement on LineMod S. Hinterstoisser et al. (2012) and Occlusion Brachmann et al. (2014a) dataset, outperforming most existing 6D pose estimation methods without using refinement.
估计已知物体的6D姿态是许多机器人和AR应用的重要任务。6D姿态估计的最新趋势是利用深度网络直接从图像中回归姿态,或者首先预测3D关键点的2D位置,然后根据关键点的2D-3D坐标关系使用PnP算法恢复姿态。然而,这些方法大多容易闭塞。在本文中,我们提出了一种简单而有效的方法,基于两阶段检测骨干(即Faster R-CNN Ren et al.(2015)),使用基于锚点的角点检测来估计物体的6D姿态。我们不是直接预测投影的三维边界框(即角)的二维坐标,而是根据对应面的二维锚点和对角线回归顶角的相对坐标。利用面对角线的几何约束进一步稳健地推断底角。实验表明,我们的方法在LineMod S. Hinterstoisser等人(2012)和Occlusion Brachmann等人(2014a)数据集上取得了显著的改进,在不使用改进的情况下优于大多数现有的6D姿态估计方法。