Conference on Robot Learning最新文献

筛选
英文 中文
MResT: Multi-Resolution Sensing for Real-Time Control with Vision-Language Models MResT:利用视觉语言模型实现实时控制的多分辨率传感技术
Conference on Robot Learning Pub Date : 2024-01-25 DOI: 10.48550/arXiv.2401.14502
Saumya Saxena, Mohit Sharma, Oliver Kroemer
{"title":"MResT: Multi-Resolution Sensing for Real-Time Control with Vision-Language Models","authors":"Saumya Saxena, Mohit Sharma, Oliver Kroemer","doi":"10.48550/arXiv.2401.14502","DOIUrl":"https://doi.org/10.48550/arXiv.2401.14502","url":null,"abstract":"Leveraging sensing modalities across diverse spatial and temporal resolutions can improve performance of robotic manipulation tasks. Multi-spatial resolution sensing provides hierarchical information captured at different spatial scales and enables both coarse and precise motions. Simultaneously multi-temporal resolution sensing enables the agent to exhibit high reactivity and real-time control. In this work, we propose a framework, MResT (Multi-Resolution Transformer), for learning generalizable language-conditioned multi-task policies that utilize sensing at different spatial and temporal resolutions using networks of varying capacities to effectively perform real time control of precise and reactive tasks. We leverage off-the-shelf pretrained vision-language models to operate on low-frequency global features along with small non-pretrained models to adapt to high frequency local feedback. Through extensive experiments in 3 domains (coarse, precise and dynamic manipulation tasks), we show that our approach significantly improves (2X on average) over recent multi-task baselines. Further, our approach generalizes well to visual and geometric variations in target objects and to varying interaction forces.","PeriodicalId":273870,"journal":{"name":"Conference on Robot Learning","volume":"299 3","pages":"2210-2228"},"PeriodicalIF":0.0,"publicationDate":"2024-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140495172","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Lidar Line Selection with Spatially-Aware Shapley Value for Cost-Efficient Depth Completion 具有空间感知的Shapley值的激光雷达选线,用于经济高效的深度完成
Conference on Robot Learning Pub Date : 2023-03-21 DOI: 10.48550/arXiv.2303.11720
Kamil Adamczewski, Christos Sakaridis, Vaishakh Patil, L. Gool
{"title":"Lidar Line Selection with Spatially-Aware Shapley Value for Cost-Efficient Depth Completion","authors":"Kamil Adamczewski, Christos Sakaridis, Vaishakh Patil, L. Gool","doi":"10.48550/arXiv.2303.11720","DOIUrl":"https://doi.org/10.48550/arXiv.2303.11720","url":null,"abstract":"Lidar is a vital sensor for estimating the depth of a scene. Typical spinning lidars emit pulses arranged in several horizontal lines and the monetary cost of the sensor increases with the number of these lines. In this work, we present the new problem of optimizing the positioning of lidar lines to find the most effective configuration for the depth completion task. We propose a solution to reduce the number of lines while retaining the up-to-the-mark quality of depth completion. Our method consists of two components, (1) line selection based on the marginal contribution of a line computed via the Shapley value and (2) incorporating line position spread to take into account its need to arrive at image-wide depth completion. Spatially-aware Shapley values (SaS) succeed in selecting line subsets that yield a depth accuracy comparable to the full lidar input while using just half of the lines.","PeriodicalId":273870,"journal":{"name":"Conference on Robot Learning","volume":"170 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117280268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Safe Robot Learning in Assistive Devices through Neural Network Repair 基于神经网络修复的辅助设备安全机器人学习
Conference on Robot Learning Pub Date : 2023-03-08 DOI: 10.48550/arXiv.2303.04431
K. Majd, Geoffrey Clark, Tanmay Khandait, Siyu Zhou, S. Sankaranarayanan, Georgios Fainekos, H. B. Amor
{"title":"Safe Robot Learning in Assistive Devices through Neural Network Repair","authors":"K. Majd, Geoffrey Clark, Tanmay Khandait, Siyu Zhou, S. Sankaranarayanan, Georgios Fainekos, H. B. Amor","doi":"10.48550/arXiv.2303.04431","DOIUrl":"https://doi.org/10.48550/arXiv.2303.04431","url":null,"abstract":"Assistive robotic devices are a particularly promising field of application for neural networks (NN) due to the need for personalization and hard-to-model human-machine interaction dynamics. However, NN based estimators and controllers may produce potentially unsafe outputs over previously unseen data points. In this paper, we introduce an algorithm for updating NN control policies to satisfy a given set of formal safety constraints, while also optimizing the original loss function. Given a set of mixed-integer linear constraints, we define the NN repair problem as a Mixed Integer Quadratic Program (MIQP). In extensive experiments, we demonstrate the efficacy of our repair method in generating safe policies for a lower-leg prosthesis.","PeriodicalId":273870,"journal":{"name":"Conference on Robot Learning","volume":"203 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115027916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
COACH: Cooperative Robot Teaching 教练:合作机器人教学
Conference on Robot Learning Pub Date : 2023-02-13 DOI: 10.48550/arXiv.2302.06199
Cunjun Yu, Yiqing Xu, Linfeng Li, David Hsu
{"title":"COACH: Cooperative Robot Teaching","authors":"Cunjun Yu, Yiqing Xu, Linfeng Li, David Hsu","doi":"10.48550/arXiv.2302.06199","DOIUrl":"https://doi.org/10.48550/arXiv.2302.06199","url":null,"abstract":"Knowledge and skills can transfer from human teachers to human students. However, such direct transfer is often not scalable for physical tasks, as they require one-to-one interaction, and human teachers are not available in sufficient numbers. Machine learning enables robots to become experts and play the role of teachers to help in this situation. In this work, we formalize cooperative robot teaching as a Markov game, consisting of four key elements: the target task, the student model, the teacher model, and the interactive teaching-learning process. Under a moderate assumption, the Markov game reduces to a partially observable Markov decision process, with an efficient approximate solution. We illustrate our approach on two cooperative tasks, one in a simulated video game and one with a real robot.","PeriodicalId":273870,"journal":{"name":"Conference on Robot Learning","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133480628","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Learning Goal-Conditioned Policies Offline with Self-Supervised Reward Shaping 基于自监督奖励塑造的目标条件策略离线学习
Conference on Robot Learning Pub Date : 2023-01-05 DOI: 10.48550/arXiv.2301.02099
Lina Mezghani, Sainbayar Sukhbaatar, Piotr Bojanowski, A. Lazaric, Alahari Karteek
{"title":"Learning Goal-Conditioned Policies Offline with Self-Supervised Reward Shaping","authors":"Lina Mezghani, Sainbayar Sukhbaatar, Piotr Bojanowski, A. Lazaric, Alahari Karteek","doi":"10.48550/arXiv.2301.02099","DOIUrl":"https://doi.org/10.48550/arXiv.2301.02099","url":null,"abstract":"Developing agents that can execute multiple skills by learning from pre-collected datasets is an important problem in robotics, where online interaction with the environment is extremely time-consuming. Moreover, manually designing reward functions for every single desired skill is prohibitive. Prior works targeted these challenges by learning goal-conditioned policies from offline datasets without manually specified rewards, through hindsight relabelling. These methods suffer from the issue of sparsity of rewards, and fail at long-horizon tasks. In this work, we propose a novel self-supervised learning phase on the pre-collected dataset to understand the structure and the dynamics of the model, and shape a dense reward function for learning policies offline. We evaluate our method on three continuous control tasks, and show that our model significantly outperforms existing approaches, especially on tasks that involve long-term planning.","PeriodicalId":273870,"journal":{"name":"Conference on Robot Learning","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129411361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Learning Road Scene-level Representations via Semantic Region Prediction 通过语义区域预测学习道路场景级表示
Conference on Robot Learning Pub Date : 2023-01-02 DOI: 10.48550/arXiv.2301.00714
Zihao Xiao, A. Yuille, Yi-Ting Chen
{"title":"Learning Road Scene-level Representations via Semantic Region Prediction","authors":"Zihao Xiao, A. Yuille, Yi-Ting Chen","doi":"10.48550/arXiv.2301.00714","DOIUrl":"https://doi.org/10.48550/arXiv.2301.00714","url":null,"abstract":"In this work, we tackle two vital tasks in automated driving systems, i.e., driver intent prediction and risk object identification from egocentric images. Mainly, we investigate the question: what would be good road scene-level representations for these two tasks? We contend that a scene-level representation must capture higher-level semantic and geometric representations of traffic scenes around ego-vehicle while performing actions to their destinations. To this end, we introduce the representation of semantic regions, which are areas where ego-vehicles visit while taking an afforded action (e.g., left-turn at 4-way intersections). We propose to learn scene-level representations via a novel semantic region prediction task and an automatic semantic region labeling algorithm. Extensive evaluations are conducted on the HDD and nuScenes datasets, and the learned representations lead to state-of-the-art performance for driver intention prediction and risk object identification.","PeriodicalId":273870,"journal":{"name":"Conference on Robot Learning","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115113805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Offline Reinforcement Learning for Visual Navigation 视觉导航的离线强化学习
Conference on Robot Learning Pub Date : 2022-12-16 DOI: 10.48550/arXiv.2212.08244
Dhruv Shah, Arjun Bhorkar, Hrish Leen, Ilya Kostrikov, Nicholas Rhinehart, S. Levine
{"title":"Offline Reinforcement Learning for Visual Navigation","authors":"Dhruv Shah, Arjun Bhorkar, Hrish Leen, Ilya Kostrikov, Nicholas Rhinehart, S. Levine","doi":"10.48550/arXiv.2212.08244","DOIUrl":"https://doi.org/10.48550/arXiv.2212.08244","url":null,"abstract":"Reinforcement learning can enable robots to navigate to distant goals while optimizing user-specified reward functions, including preferences for following lanes, staying on paved paths, or avoiding freshly mowed grass. However, online learning from trial-and-error for real-world robots is logistically challenging, and methods that instead can utilize existing datasets of robotic navigation data could be significantly more scalable and enable broader generalization. In this paper, we present ReViND, the first offline RL system for robotic navigation that can leverage previously collected data to optimize user-specified reward functions in the real-world. We evaluate our system for off-road navigation without any additional data collection or fine-tuning, and show that it can navigate to distant goals using only offline training from this dataset, and exhibit behaviors that qualitatively differ based on the user-specified reward function.","PeriodicalId":273870,"journal":{"name":"Conference on Robot Learning","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121279339","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
JFP: Joint Future Prediction with Interactive Multi-Agent Modeling for Autonomous Driving 基于交互式多智能体模型的自动驾驶联合未来预测
Conference on Robot Learning Pub Date : 2022-12-16 DOI: 10.48550/arXiv.2212.08710
Wenjie Luo, C. Park, Andre Cornman, Benjamin Sapp, Drago Anguelov
{"title":"JFP: Joint Future Prediction with Interactive Multi-Agent Modeling for Autonomous Driving","authors":"Wenjie Luo, C. Park, Andre Cornman, Benjamin Sapp, Drago Anguelov","doi":"10.48550/arXiv.2212.08710","DOIUrl":"https://doi.org/10.48550/arXiv.2212.08710","url":null,"abstract":"We propose JFP, a Joint Future Prediction model that can learn to generate accurate and consistent multi-agent future trajectories. For this task, many different methods have been proposed to capture social interactions in the encoding part of the model, however, considerably less focus has been placed on representing interactions in the decoder and output stages. As a result, the predicted trajectories are not necessarily consistent with each other, and often result in unrealistic trajectory overlaps. In contrast, we propose an end-to-end trainable model that learns directly the interaction between pairs of agents in a structured, graphical model formulation in order to generate consistent future trajectories. It sets new state-of-the-art results on Waymo Open Motion Dataset (WOMD) for the interactive setting. We also investigate a more complex multi-agent setting for both WOMD and a larger internal dataset, where our approach improves significantly on the trajectory overlap metrics while obtaining on-par or better performance on single-agent trajectory metrics.","PeriodicalId":273870,"journal":{"name":"Conference on Robot Learning","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125621228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Learning Markerless Robot-Depth Camera Calibration and End-Effector Pose Estimation 学习无标记机器人深度摄像机标定和末端执行器姿态估计
Conference on Robot Learning Pub Date : 2022-12-15 DOI: 10.48550/arXiv.2212.07567
B. C. Sefercik, Barış Akgün
{"title":"Learning Markerless Robot-Depth Camera Calibration and End-Effector Pose Estimation","authors":"B. C. Sefercik, Barış Akgün","doi":"10.48550/arXiv.2212.07567","DOIUrl":"https://doi.org/10.48550/arXiv.2212.07567","url":null,"abstract":"Traditional approaches to extrinsic calibration use fiducial markers and learning-based approaches rely heavily on simulation data. In this work, we present a learning-based markerless extrinsic calibration system that uses a depth camera and does not rely on simulation data. We learn models for end-effector (EE) segmentation, single-frame rotation prediction and keypoint detection, from automatically generated real-world data. We use a transformation trick to get EE pose estimates from rotation predictions and a matching algorithm to get EE pose estimates from keypoint predictions. We further utilize the iterative closest point algorithm, multiple-frames, filtering and outlier detection to increase calibration robustness. Our evaluations with training data from multiple camera poses and test data from previously unseen poses give sub-centimeter and sub-deciradian average calibration and pose estimation errors. We also show that a carefully selected single training pose gives comparable results.","PeriodicalId":273870,"journal":{"name":"Conference on Robot Learning","volume":"2 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116862241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
HUM3DIL: Semi-supervised Multi-modal 3D Human Pose Estimation for Autonomous Driving HUM3DIL:用于自动驾驶的半监督多模态三维人体姿态估计
Conference on Robot Learning Pub Date : 2022-12-15 DOI: 10.48550/arXiv.2212.07729
Andrei Zanfir, M. Zanfir, Alexander N. Gorban, Jingwei Ji, Yin Zhou, Drago Anguelov, C. Sminchisescu
{"title":"HUM3DIL: Semi-supervised Multi-modal 3D Human Pose Estimation for Autonomous Driving","authors":"Andrei Zanfir, M. Zanfir, Alexander N. Gorban, Jingwei Ji, Yin Zhou, Drago Anguelov, C. Sminchisescu","doi":"10.48550/arXiv.2212.07729","DOIUrl":"https://doi.org/10.48550/arXiv.2212.07729","url":null,"abstract":"Autonomous driving is an exciting new industry, posing important research questions. Within the perception module, 3D human pose estimation is an emerging technology, which can enable the autonomous vehicle to perceive and understand the subtle and complex behaviors of pedestrians. While hardware systems and sensors have dramatically improved over the decades -- with cars potentially boasting complex LiDAR and vision systems and with a growing expansion of the available body of dedicated datasets for this newly available information -- not much work has been done to harness these novel signals for the core problem of 3D human pose estimation. Our method, which we coin HUM3DIL (HUMan 3D from Images and LiDAR), efficiently makes use of these complementary signals, in a semi-supervised fashion and outperforms existing methods with a large margin. It is a fast and compact model for onboard deployment. Specifically, we embed LiDAR points into pixel-aligned multi-modal features, which we pass through a sequence of Transformer refinement stages. Quantitative experiments on the Waymo Open Dataset support these claims, where we achieve state-of-the-art results on the task of 3D pose estimation.","PeriodicalId":273870,"journal":{"name":"Conference on Robot Learning","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125939427","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信