CUDA accelerated robot localization and mapping

H. Zhang, F. Martin
{"title":"CUDA accelerated robot localization and mapping","authors":"H. Zhang, F. Martin","doi":"10.1109/TePRA.2013.6556350","DOIUrl":null,"url":null,"abstract":"We present a method to accelerate robot localization and mapping by using CUDA (Compute Unified Device Architecture), the general purpose parallel computing platform on NVIDIA GPUs. In robotics, the particle filter-based SLAM (Simultaneous Localization and Mapping) algorithm has many applications, but is computationally intensive. Prior work has used CUDA to accelerate various robot applications, but particle filter-based SLAM has not been implemented on CUDA yet. Because computations on the particles are independent of each other in this algorithm, CUDA acceleration should be highly effective. We have implemented the SLAM algorithm's most time consuming step, particle weight calculation, and optimized memory access by using texture memory to alleviate memory bottleneck and fully leverage the parallel processing power. Our experiments have shown the performance has increased by an order of magnitude or more. The results indicate that oftloading to GPU is a cost-effective way to improve SLAM algorithm performance.","PeriodicalId":102284,"journal":{"name":"2013 IEEE Conference on Technologies for Practical Robot Applications (TePRA)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"23","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 IEEE Conference on Technologies for Practical Robot Applications (TePRA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/TePRA.2013.6556350","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 23

Abstract

We present a method to accelerate robot localization and mapping by using CUDA (Compute Unified Device Architecture), the general purpose parallel computing platform on NVIDIA GPUs. In robotics, the particle filter-based SLAM (Simultaneous Localization and Mapping) algorithm has many applications, but is computationally intensive. Prior work has used CUDA to accelerate various robot applications, but particle filter-based SLAM has not been implemented on CUDA yet. Because computations on the particles are independent of each other in this algorithm, CUDA acceleration should be highly effective. We have implemented the SLAM algorithm's most time consuming step, particle weight calculation, and optimized memory access by using texture memory to alleviate memory bottleneck and fully leverage the parallel processing power. Our experiments have shown the performance has increased by an order of magnitude or more. The results indicate that oftloading to GPU is a cost-effective way to improve SLAM algorithm performance.
CUDA加速机器人定位和绘图
我们提出了一种利用基于NVIDIA gpu的通用并行计算平台CUDA(计算统一设备架构)来加速机器人定位和映射的方法。在机器人技术中,基于粒子滤波的SLAM (Simultaneous Localization and Mapping)算法应用广泛,但计算量较大。之前的工作已经使用CUDA来加速各种机器人应用,但是基于粒子滤波的SLAM还没有在CUDA上实现。由于该算法中对粒子的计算是相互独立的,因此CUDA加速应该是非常有效的。我们实现了SLAM算法中最耗时的步骤——粒子权值计算,并利用纹理存储器优化了内存访问,缓解了内存瓶颈,充分利用了并行处理能力。我们的实验表明,性能提高了一个数量级或更多。结果表明,将算法卸载到GPU上是提高SLAM算法性能的一种经济有效的方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信