Yawen Lu, Sophia Kourian, C. Salvaggio, Chenliang Xu, G. Lu
{"title":"增强现实单图像三维车辆姿态估计","authors":"Yawen Lu, Sophia Kourian, C. Salvaggio, Chenliang Xu, G. Lu","doi":"10.1109/GlobalSIP45357.2019.8969201","DOIUrl":null,"url":null,"abstract":"The intent of this paper is to introduce a novel method for 3D vehicle pose estimation, a critical component of augmented reality. The proposed method is able to recover the location of a specific object from a single image by combining pretrained reliable semantic segmentation and improved single image depth estimation. Our method exploits a novel pose estimation technique by generating new 2D images created from the projections of rotated point clouds. The rotation of the specific object is able to be predicted. Augmented objects can be shifted, rotated and scaled correctly based on the recovered orientation and localization. Through accurate vehicle pose estimation, virtual vehicles are able to be augmented accurately in place of real vehicles. The effectiveness of our method is verified by comparison with other recent pose estimation methods on the challenging KITTI 3D benchmark. Further experiments on the Cityscapes dataset also demonstrates good robustness in the method. Without requiring ground truth 3D vehicle pose labels for training, our model is able to produce competitive and robust performance in 3D vehicle pose estimation.","PeriodicalId":221378,"journal":{"name":"2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Single Image 3D Vehicle Pose Estimation for Augmented Reality\",\"authors\":\"Yawen Lu, Sophia Kourian, C. Salvaggio, Chenliang Xu, G. Lu\",\"doi\":\"10.1109/GlobalSIP45357.2019.8969201\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The intent of this paper is to introduce a novel method for 3D vehicle pose estimation, a critical component of augmented reality. The proposed method is able to recover the location of a specific object from a single image by combining pretrained reliable semantic segmentation and improved single image depth estimation. Our method exploits a novel pose estimation technique by generating new 2D images created from the projections of rotated point clouds. The rotation of the specific object is able to be predicted. Augmented objects can be shifted, rotated and scaled correctly based on the recovered orientation and localization. Through accurate vehicle pose estimation, virtual vehicles are able to be augmented accurately in place of real vehicles. The effectiveness of our method is verified by comparison with other recent pose estimation methods on the challenging KITTI 3D benchmark. Further experiments on the Cityscapes dataset also demonstrates good robustness in the method. Without requiring ground truth 3D vehicle pose labels for training, our model is able to produce competitive and robust performance in 3D vehicle pose estimation.\",\"PeriodicalId\":221378,\"journal\":{\"name\":\"2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP)\",\"volume\":\"13 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/GlobalSIP45357.2019.8969201\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Global Conference on Signal and Information Processing (GlobalSIP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/GlobalSIP45357.2019.8969201","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Single Image 3D Vehicle Pose Estimation for Augmented Reality
The intent of this paper is to introduce a novel method for 3D vehicle pose estimation, a critical component of augmented reality. The proposed method is able to recover the location of a specific object from a single image by combining pretrained reliable semantic segmentation and improved single image depth estimation. Our method exploits a novel pose estimation technique by generating new 2D images created from the projections of rotated point clouds. The rotation of the specific object is able to be predicted. Augmented objects can be shifted, rotated and scaled correctly based on the recovered orientation and localization. Through accurate vehicle pose estimation, virtual vehicles are able to be augmented accurately in place of real vehicles. The effectiveness of our method is verified by comparison with other recent pose estimation methods on the challenging KITTI 3D benchmark. Further experiments on the Cityscapes dataset also demonstrates good robustness in the method. Without requiring ground truth 3D vehicle pose labels for training, our model is able to produce competitive and robust performance in 3D vehicle pose estimation.