Maxime Chareyre;Pierre Fournier;Julien Moras;Jean-Marc Bourinet;Youcef Mezouar
{"title":"Unsupervised Discovery of Objects Physical Properties Through Maximum Entropy Reinforcement Learning","authors":"Maxime Chareyre;Pierre Fournier;Julien Moras;Jean-Marc Bourinet;Youcef Mezouar","doi":"10.1109/LRA.2025.3545382","DOIUrl":null,"url":null,"abstract":"Understanding the environment is crucial for autonomous robots to perform navigation and manipulation tasks. Never-seen-before objects may have complex appearances and dynamics, where only physical interactions can help to identify visually hidden properties like mass or friction. In this work we propose a baseline for the newly defined problem of using physical interactions to discover unknown properties of objects, without prior knowledge of them or any supervision. The agent first uses intrinsically motivated unsupervised reinforcement learning to learn how to interact with objects, so as to get observations with a level of information which eases the physical properties estimation. A self-supervised predictive task is then set up while following the learned behaviour to extract a latent representation of the physical properties of an object. When applied to a simulated mobile robot in presence of varying objects, the proposed baseline identifies and differentiates categorical properties, e.g. shape, and quantifies continuous properties, e.g. mass and friction, with excellent correlations to their true values even from noisy observations. It achieves significantly better results than simple interactions of a policy that performs poor exploration. This work provides an implementation of a functional, object-oriented action-perception cycle for embodied robotic agents.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 4","pages":"3723-3730"},"PeriodicalIF":4.6000,"publicationDate":"2025-02-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Robotics and Automation Letters","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10902017/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Understanding the environment is crucial for autonomous robots to perform navigation and manipulation tasks. Never-seen-before objects may have complex appearances and dynamics, where only physical interactions can help to identify visually hidden properties like mass or friction. In this work we propose a baseline for the newly defined problem of using physical interactions to discover unknown properties of objects, without prior knowledge of them or any supervision. The agent first uses intrinsically motivated unsupervised reinforcement learning to learn how to interact with objects, so as to get observations with a level of information which eases the physical properties estimation. A self-supervised predictive task is then set up while following the learned behaviour to extract a latent representation of the physical properties of an object. When applied to a simulated mobile robot in presence of varying objects, the proposed baseline identifies and differentiates categorical properties, e.g. shape, and quantifies continuous properties, e.g. mass and friction, with excellent correlations to their true values even from noisy observations. It achieves significantly better results than simple interactions of a policy that performs poor exploration. This work provides an implementation of a functional, object-oriented action-perception cycle for embodied robotic agents.
期刊介绍:
The scope of this journal is to publish peer-reviewed articles that provide a timely and concise account of innovative research ideas and application results, reporting significant theoretical findings and application case studies in areas of robotics and automation.