{"title":"机器人视觉旋转不变目标识别","authors":"C. Vilar, Silvia Krug, B. Thornberg","doi":"10.1145/3365265.3365273","DOIUrl":null,"url":null,"abstract":"Depth cameras have enhanced the environment perception for robotic applications significantly. They allow to measure true distances and thus enable a 3D measurement of the robot surroundings. In order to enable robust robot vision, the objects recognition has to handle rotated data because object can be viewed from different dynamic perspectives when the robot is moving. Therefore, the 3D descriptors used of object recognition for robotic applications have to be rotation invariant and implementable on the embedded system, with limited memory and computing resources. With the popularization of the depth cameras, the Histogram of Gradients (HOG) descriptor has been extended to recognize also 3D volumetric objects (3DVHOG). Unfortunately, both version are not rotation invariant. There are different methods to achieve rotation invariance for 3DVHOG, but they increase significantly the computational cost of the overall data processing. Hence, they are unfeasible to be implemented in a low cost processor for real-time operation. In this paper, we propose an object pose normalization method to achieve 3DVHOG rotation invariance while reducing the number of processing operations as much as possible. Our method is based on Principal Component Analysis (PCA) normalization. We tested our method using the Princeton Modelnet10 dataset.","PeriodicalId":358714,"journal":{"name":"Proceedings of the 2019 3rd International Conference on Automation, Control and Robots","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Rotational Invariant Object Recognition for Robotic Vision\",\"authors\":\"C. Vilar, Silvia Krug, B. Thornberg\",\"doi\":\"10.1145/3365265.3365273\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Depth cameras have enhanced the environment perception for robotic applications significantly. They allow to measure true distances and thus enable a 3D measurement of the robot surroundings. In order to enable robust robot vision, the objects recognition has to handle rotated data because object can be viewed from different dynamic perspectives when the robot is moving. Therefore, the 3D descriptors used of object recognition for robotic applications have to be rotation invariant and implementable on the embedded system, with limited memory and computing resources. With the popularization of the depth cameras, the Histogram of Gradients (HOG) descriptor has been extended to recognize also 3D volumetric objects (3DVHOG). Unfortunately, both version are not rotation invariant. There are different methods to achieve rotation invariance for 3DVHOG, but they increase significantly the computational cost of the overall data processing. Hence, they are unfeasible to be implemented in a low cost processor for real-time operation. In this paper, we propose an object pose normalization method to achieve 3DVHOG rotation invariance while reducing the number of processing operations as much as possible. Our method is based on Principal Component Analysis (PCA) normalization. We tested our method using the Princeton Modelnet10 dataset.\",\"PeriodicalId\":358714,\"journal\":{\"name\":\"Proceedings of the 2019 3rd International Conference on Automation, Control and Robots\",\"volume\":\"5 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-10-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2019 3rd International Conference on Automation, Control and Robots\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3365265.3365273\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2019 3rd International Conference on Automation, Control and Robots","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3365265.3365273","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
摘要
深度相机极大地增强了机器人应用的环境感知能力。它们可以测量真实距离,从而实现对机器人周围环境的3D测量。为了实现机器人视觉的鲁棒性,物体识别必须处理旋转数据,因为当机器人移动时,物体可以从不同的动态角度观察。因此,机器人应用中用于物体识别的三维描述符必须是旋转不变性的,并且可以在内存和计算资源有限的嵌入式系统上实现。随着深度相机的普及,梯度直方图(Histogram of Gradients, HOG)描述符也被扩展到三维体积目标(3dvolumetric objects, 3DVHOG)的识别。不幸的是,这两个版本都不是旋转不变的。实现3DVHOG旋转不变性的方法有很多,但都会显著增加整个数据处理的计算成本。因此,在低成本的处理器上实现实时操作是不可行的。在本文中,我们提出了一种目标位姿归一化方法,在尽可能减少处理操作次数的同时实现3DVHOG旋转不变性。我们的方法是基于主成分分析(PCA)归一化。我们使用普林斯顿Modelnet10数据集测试了我们的方法。
Rotational Invariant Object Recognition for Robotic Vision
Depth cameras have enhanced the environment perception for robotic applications significantly. They allow to measure true distances and thus enable a 3D measurement of the robot surroundings. In order to enable robust robot vision, the objects recognition has to handle rotated data because object can be viewed from different dynamic perspectives when the robot is moving. Therefore, the 3D descriptors used of object recognition for robotic applications have to be rotation invariant and implementable on the embedded system, with limited memory and computing resources. With the popularization of the depth cameras, the Histogram of Gradients (HOG) descriptor has been extended to recognize also 3D volumetric objects (3DVHOG). Unfortunately, both version are not rotation invariant. There are different methods to achieve rotation invariance for 3DVHOG, but they increase significantly the computational cost of the overall data processing. Hence, they are unfeasible to be implemented in a low cost processor for real-time operation. In this paper, we propose an object pose normalization method to achieve 3DVHOG rotation invariance while reducing the number of processing operations as much as possible. Our method is based on Principal Component Analysis (PCA) normalization. We tested our method using the Princeton Modelnet10 dataset.