{"title":"机器人代理感和物体持久性的深度卷积神经网络模型","authors":"Claus Lang, G. Schillaci, V. Hafner","doi":"10.1109/DEVLRN.2018.8761015","DOIUrl":null,"url":null,"abstract":"This work investigates the role of predictive models in the implementation of basic cognitive skills in robots, such as the capability to distinguish between self-generated actions and those generated by other individuals and the capability to maintain an enhanced internal visual representation of the world, where objects covered by the robot's own body in the original image may be visible in the enhanced one. A developmental approach is adopted for this purpose. In particular, a humanoid robot is learning, through a self-exploration behaviour, the sensory consequences (in the visual domain) of self-generated movements. The generated sensorimotor experience is used as training data for a deep convolutional neural network that maps proprioceptive and motor data (e.g. initial arm joint positions and applied motor commands) onto the visual consequences of these actions. This forward model is then used in two experiments. First, for generating visual predictions of self-generated movements, which are compared to actual visual perceptions and then used to compute a prediction error. This error is shown to be higher when there is an external subject performing actions, compared to situations where the robot is observing only itself. This supports the idea that prediction errors may serve as a cue for distinguishing between self and other, a fundamental prerequisite for the sense of agency. Secondly, we show how predictions can be used to attenuate self-generated movements, and thus create enhanced visual perceptions, where the sight of objects - originally occluded by the robot body - is still maintained. This may represent an important tool both for cognitive development in robots and for the understanding of the sense of object permanence in humans.","PeriodicalId":236346,"journal":{"name":"2018 Joint IEEE 8th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"19","resultStr":"{\"title\":\"A Deep Convolutional Neural Network Model for Sense of Agency and Object Permanence in Robots\",\"authors\":\"Claus Lang, G. Schillaci, V. Hafner\",\"doi\":\"10.1109/DEVLRN.2018.8761015\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This work investigates the role of predictive models in the implementation of basic cognitive skills in robots, such as the capability to distinguish between self-generated actions and those generated by other individuals and the capability to maintain an enhanced internal visual representation of the world, where objects covered by the robot's own body in the original image may be visible in the enhanced one. A developmental approach is adopted for this purpose. In particular, a humanoid robot is learning, through a self-exploration behaviour, the sensory consequences (in the visual domain) of self-generated movements. The generated sensorimotor experience is used as training data for a deep convolutional neural network that maps proprioceptive and motor data (e.g. initial arm joint positions and applied motor commands) onto the visual consequences of these actions. This forward model is then used in two experiments. First, for generating visual predictions of self-generated movements, which are compared to actual visual perceptions and then used to compute a prediction error. This error is shown to be higher when there is an external subject performing actions, compared to situations where the robot is observing only itself. This supports the idea that prediction errors may serve as a cue for distinguishing between self and other, a fundamental prerequisite for the sense of agency. Secondly, we show how predictions can be used to attenuate self-generated movements, and thus create enhanced visual perceptions, where the sight of objects - originally occluded by the robot body - is still maintained. This may represent an important tool both for cognitive development in robots and for the understanding of the sense of object permanence in humans.\",\"PeriodicalId\":236346,\"journal\":{\"name\":\"2018 Joint IEEE 8th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob)\",\"volume\":\"36 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"19\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 Joint IEEE 8th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/DEVLRN.2018.8761015\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 Joint IEEE 8th International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DEVLRN.2018.8761015","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Deep Convolutional Neural Network Model for Sense of Agency and Object Permanence in Robots
This work investigates the role of predictive models in the implementation of basic cognitive skills in robots, such as the capability to distinguish between self-generated actions and those generated by other individuals and the capability to maintain an enhanced internal visual representation of the world, where objects covered by the robot's own body in the original image may be visible in the enhanced one. A developmental approach is adopted for this purpose. In particular, a humanoid robot is learning, through a self-exploration behaviour, the sensory consequences (in the visual domain) of self-generated movements. The generated sensorimotor experience is used as training data for a deep convolutional neural network that maps proprioceptive and motor data (e.g. initial arm joint positions and applied motor commands) onto the visual consequences of these actions. This forward model is then used in two experiments. First, for generating visual predictions of self-generated movements, which are compared to actual visual perceptions and then used to compute a prediction error. This error is shown to be higher when there is an external subject performing actions, compared to situations where the robot is observing only itself. This supports the idea that prediction errors may serve as a cue for distinguishing between self and other, a fundamental prerequisite for the sense of agency. Secondly, we show how predictions can be used to attenuate self-generated movements, and thus create enhanced visual perceptions, where the sight of objects - originally occluded by the robot body - is still maintained. This may represent an important tool both for cognitive development in robots and for the understanding of the sense of object permanence in humans.