Alexandre Devillers, Valentin Chaffraix, Frederic Armetta, S. Duffner, Mathieu Lefort
{"title":"The Impact of Action in Visual Representation Learning","authors":"Alexandre Devillers, Valentin Chaffraix, Frederic Armetta, S. Duffner, Mathieu Lefort","doi":"10.1109/ICDL53763.2022.9962210","DOIUrl":"https://doi.org/10.1109/ICDL53763.2022.9962210","url":null,"abstract":"Sensori-motor theories, inspired by work in neuroscience, psychology and cognitive science, claim that actions, through learning and mastering of a predictive model, are a key element in the perception of the environment. On the computational side, in the domains of representation learning and reinforcement learning, models are increasingly using self-supervised pretext tasks, such as predictive or contrastive ones, in order to increase the performance on their main task. These pretext tasks are action-related even if the action itself is usually not used in the model. In this paper, we propose to study the influence of considering action in the learning of visual representations in deep neural network models, an aspect which is often underestimated w.r.t. sensori-motor theories. More precisely, we quantity two independent factors: 1-whether or not to use the action during the learning of visual characteristics, and 2-whether or not to integrate the action in the representations of the current images. Other aspects will be kept as simple and comparable as possible, that is why we will not consider any specific action policies and combine simple architectures (VAE and LSTM), while using datasets derived from MNIST. In this context, our results show that explicitly including action in the learning process and in the representations improves the performance of the model, which opens interesting perspectives to improve state-of-the-art models of representation learning.","PeriodicalId":274171,"journal":{"name":"2022 IEEE International Conference on Development and Learning (ICDL)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130807859","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ørjan Strand, Didrik Spanne Reilstad, Zhenying Wu, Bruno C. da Silva, J. Tørresen, K. Ellefsen
{"title":"RADAR: Reactive and Deliberative Adaptive Reasoning - Learning When to Think Fast and When to Think Slow","authors":"Ørjan Strand, Didrik Spanne Reilstad, Zhenying Wu, Bruno C. da Silva, J. Tørresen, K. Ellefsen","doi":"10.1109/ICDL53763.2022.9962202","DOIUrl":"https://doi.org/10.1109/ICDL53763.2022.9962202","url":null,"abstract":"When designing and deploying Reinforcement Learning (RL) algorithms, one typically selects a single value for the discount rate, which results in an agent that will always be equally reactive or deliberative. However, similarly to humans, RL agents can benefit from adapting their planning horizon to the current context. To enable this, we propose a novel algorithm: RADAR: Reactive and Deliberate Adaptive Reasoning. RADAR enables an agent to learn to adaptively choose a level of deliberation and reactivity according to the state it is in, given that there are cases where one mode of operation is better than the other. Through experiments in a grid world, we verify that the RADAR agent has the capability to adapt its reasoning modality to the current context. In addition, we observe that the RADAR agent exhibits different preferences regarding its thinking modes when a penalty for mental effort is included in its mathematical formulation.","PeriodicalId":274171,"journal":{"name":"2022 IEEE International Conference on Development and Learning (ICDL)","volume":"2015 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127676413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"What Kind of Player are You? Continuous Learning of a Player Profile for Adaptive Robot Teleoperation","authors":"Mélanie Jouaiti, K. Dautenhahn","doi":"10.1109/ICDL53763.2022.9962211","DOIUrl":"https://doi.org/10.1109/ICDL53763.2022.9962211","url":null,"abstract":"Play is important for child development and robot-assisted play is very popular in Human-Robot Interaction as it creates more engaging and realistic setups for user studies. Adaptive game-play is also an emerging research field and a good way to provide a personalized experience while adapting to individual user’s needs. In this paper, we analyze joystick data and investigate player learning during a robot navigation game. We collected joystick data from healthy adult participants playing a game with our custom robot MyJay, while participants teleoperated the robot to perform goal-directed navigation. We evaluated the performance of both novice and proficient joystick users. Based on this analysis, we propose some robot learning mechanisms to provide a personalized game experience. Our findings can help improving human-robot interaction in the context of teleoperation in general, and could be particularly impactful for children with disabilities who have problems operating off-the-shelf joysticks.","PeriodicalId":274171,"journal":{"name":"2022 IEEE International Conference on Development and Learning (ICDL)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132457763","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Dream to Pose in a Tendon-Driven Manipulator with Muscle Synergy","authors":"Matthew Ishige, T. Taniguchi, Yoshihiro Kawahara","doi":"10.1109/ICDL53763.2022.9962220","DOIUrl":"https://doi.org/10.1109/ICDL53763.2022.9962220","url":null,"abstract":"Bio-inspired tendon-driven manipulators have the potential to achieve human-level dexterity. However, their control is more complex than prevailing robotic hands because the relation between actuation and hand motion (Jacobian) is hard to obtain. On the other hand, humans maneuver their complex hands skillfully and conduct adaptive object grasping and manipulation. We conjecture that the foundation of this ability is a visual posing of hands (i.e., a skill to make arbitrary hand poses with visual and proprioceptive feedback). Children develop this skill before or in parallel with learning grasping and manipulation. Inspired by this developmental process, this study explored a method to equip compliant tendon-driven manipulators with the visual posing. To overcome the complexity of the system, we used a learning-based approach. Specifically, we adopted PlaNet, model-based reinforcement learning that leverages a dynamics model on a compact latent representation. To further accelerate learning, we restricted the control space using the idea of muscle synergy found in the human body control. We validated the effectiveness of the proposed method in a simulation. We also demonstrated that the posing skill acquired using our method is useful for object grasping. This study will contribute to achieving human-level dexterity in manipulations.","PeriodicalId":274171,"journal":{"name":"2022 IEEE International Conference on Development and Learning (ICDL)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132939544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Paul Schydlo, Laura Santos, Atabak Dehban, A. John, J. Santos-Victor
{"title":"I Have Seen that Before: Memory Augmented Neural Network for Learning Affordances by Analogy","authors":"Paul Schydlo, Laura Santos, Atabak Dehban, A. John, J. Santos-Victor","doi":"10.1109/ICDL53763.2022.9962191","DOIUrl":"https://doi.org/10.1109/ICDL53763.2022.9962191","url":null,"abstract":"Humans show a remarkable ability to quickly adapt to new situations without forgetting past ones. Existing learning methods still face problems with catastrophic forgetting when learning about new situations and contexts while taking many iterations to adjust to new input and output pairs. In this work, we propose the application of a Memory augmented network to the problem of learning tool affordances. We consider a network that explicitly indexes an episodic memory of past experiences and retrieves samples of past experience to reason about new situations by analogy, in an approach we call affordances by analogy. The work takes advantage of a tool-object interaction dataset to learn affordances. Our experiments show the model outperforms the baselines in the low sample regime and retains information better when re-trained on a different data distribution. Hinting at a promising direction, this work could enable learning algorithms to retain information better.","PeriodicalId":274171,"journal":{"name":"2022 IEEE International Conference on Development and Learning (ICDL)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133894587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Aramis Augusto Bonzini, L. Seminara, Simone Macciò, A. Carfì, L. Jamone
{"title":"Leveraging symmetry detection to speed up haptic object exploration in robots","authors":"Aramis Augusto Bonzini, L. Seminara, Simone Macciò, A. Carfì, L. Jamone","doi":"10.1109/ICDL53763.2022.9962206","DOIUrl":"https://doi.org/10.1109/ICDL53763.2022.9962206","url":null,"abstract":"Most objects are symmetric. In fact, humans are very good at detecting symmetry, both by vision and by touch, and they use such information to facilitate the perception of other object properties, such as shape and size; overall, this contributes to human’s ability to successfully manipulate objects in unstructured environments. Inspired by this human skill, in this paper we propose a haptic exploration procedure that enables a robot to detect object symmetry, and uses such information to estimate the shape of an object with higher accuracy and in less time. We achieve this by incorporating symmetries in a Gaussian Process model, and by introducing a novel strategy to detect the presence of such symmetry. We report results obtained with a Baxter robot equipped with a custom tactile sensor on the gripper: we show that when the robot explores objects with unknown symmetries the time required to estimate the object shape is reduced by up to 50% thanks to our method.","PeriodicalId":274171,"journal":{"name":"2022 IEEE International Conference on Development and Learning (ICDL)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132882156","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Embodied Attention in Word-Object Mapping: A Developmental Cognitive Robotics Model","authors":"Luca Raggioli, A. Cangelosi","doi":"10.1109/ICDL53763.2022.9962189","DOIUrl":"https://doi.org/10.1109/ICDL53763.2022.9962189","url":null,"abstract":"Developmental Robotics models provide useful tools to study and understand the language learning process in infants and robots. These models allow us to describe key mechanisms of language development, such as statistical learning, the role of embodiment, and the impact of the attention payed to an object while learning its name. Robots can be particularly well suited for this type of problems, because they cover both a physical manipulation of the environment and mathematical modeling of the temporal changes of the learned concepts. In this work we present a computational representation of the impact of embodiment and attention on word learning, relying on sensory data collected with a real robotic agent in a real world scenario. Results show that the cognitive architecture designed for this scenario is able to capture the changes underlying the moving object in the field of view of the robot. The architecture successfully handles the temporal relationship in moving items and manages to show the effects of the embodied attention on word-object mapping.","PeriodicalId":274171,"journal":{"name":"2022 IEEE International Conference on Development and Learning (ICDL)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121889497","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Getting Priorities Right: Intrinsic Motivation with Multi-Objective Reinforcement Learning","authors":"Yusuf Al-Husaini, Matthias Rolf","doi":"10.1109/ICDL53763.2022.9962187","DOIUrl":"https://doi.org/10.1109/ICDL53763.2022.9962187","url":null,"abstract":"Intrinsic motivation is a common method to facilitate exploration in reinforcement learning agents. Curiosity is thereby supposed to aid the learning of a primary goal. However, indulging in curiosity may also stand in conflict with more urgent or essential objectives such as self-sustenance. This paper addresses the problem of balancing curiosity, and correctly prioritising other needs in a reinforcement learning context. We demonstrate the use of the multi-objective reinforcement learning framework C-MORE to integrate curiosity, and compare results to a standard linear reinforcement learning integration. Results clearly demonstrate that curiosity can be modelled with the priority-objective reinforcement learning paradigm. In particular, C-MORE is found to explore robustly while maintaining self-sustenance objectives, whereas the linear approach is found to over-explore and take unnecessary risks. The findings demonstrate a significant weakness of the common linear integration method for intrinsic motivation, and the need to acknowledge the potential conflicts between curiosity and other objectives in a multi-objective framework.","PeriodicalId":274171,"journal":{"name":"2022 IEEE International Conference on Development and Learning (ICDL)","volume":"86 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124801304","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Kristína Malinovská, I. Farkaš, Jana Harvanová, M. Hoffmann
{"title":"A connectionist model of associating proprioceptive and tactile modalities in a humanoid robot","authors":"Kristína Malinovská, I. Farkaš, Jana Harvanová, M. Hoffmann","doi":"10.1109/ICDL53763.2022.9962195","DOIUrl":"https://doi.org/10.1109/ICDL53763.2022.9962195","url":null,"abstract":"Postnatal development in infants involves building the body schema based on integrating information from different modalities. An early phase of this complex process involves coupling proprioceptive inputs with tactile information during self-touch enabled by motor babbling. Such functionality is also desirable in humanoid robots that can serve as embodied instantiation of cognitive learning. We describe a simple connectionist model composed of neural networks that learns the proprioceptive-tactile representations on a simulated iCub humanoid robot. Input signals from both modalities – joint angles and touch stimuli on both upper limbs – are first self-organized in neural maps and then connected using a universal bidirectional associative network (UBAL). The model demonstrates the ability to predict touch and its location from proprioceptive information with relatively high accuracy. We also discuss limitations of the model and the ideas for future work.","PeriodicalId":274171,"journal":{"name":"2022 IEEE International Conference on Development and Learning (ICDL)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125327168","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Juan José Gamboa-Montero, M. Basiri, J. C. Castillo, Sara Marques-Villarroya, M. Salichs
{"title":"Real-Time Acoustic Touch Localization in Human-Robot Interaction based on Steered Response Power","authors":"Juan José Gamboa-Montero, M. Basiri, J. C. Castillo, Sara Marques-Villarroya, M. Salichs","doi":"10.1109/ICDL53763.2022.9962225","DOIUrl":"https://doi.org/10.1109/ICDL53763.2022.9962225","url":null,"abstract":"The sense of touch plays an important role in Human-Robot Interaction, allowing a natural form of communication with humans and improving the rest of the robot’s senses. This is even more important in the subject of social robotics, where robots must interact with people and adhere to social conventions while doing so. Touch interfaces can be implemented as part of a robotic platform serving multiple purposes. Some examples include the use of tactile commands to control the movement of a robot, or attempting to endow a robot with the ability to understand human emotional states. This work proposes a system to localize a contact performed on the rigid, non-planar shell of a service robot in real-time, based on set of spatially separated piezo transducers attached to the shell of the robot and the Steered Response Power sound source localization algorithm. Results show the potential capability of the system to correctly detect and localize human touches.","PeriodicalId":274171,"journal":{"name":"2022 IEEE International Conference on Development and Learning (ICDL)","volume":"131 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122423099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}