{"title":"认知有效的终身学习模式","authors":"Hanne Say, E. Oztop","doi":"10.1109/ROBIO58561.2023.10355028","DOIUrl":null,"url":null,"abstract":"In continual learning, usually a sequence of tasks are given to a learning agent and the performance of the agent after learning is measured in terms of resistance to catastrophic forgetting, efficacy of knowledge transfer and overall performance on the individual tasks. On the other hand, in multi-task learning, the system is designed to simultaneously acquire knowledge in multiple tasks, often through offline batch learning. A more cognitively valid scenario for lifelong robot learning would be to have a robotic agent to autonomously decide which task to engage and disengage while leveraging many-to-many knowledge transfer ability among tasks during online learning. In this study, we propose a novel lifelong robot learning architecture to fulfill the aforementioned desiderata, and show its validity in an environment where a robot learns the effects of its actions in different task settings. To realize the proposed model, we adopt learning progress measure for task selection, and have the tasks learn by independent neural networks with special structure that allows access to the neural layers of the non-selected tasks. The experiments conducted with a simulated robot arm in an object interaction scenario show that the proposed architecture yields better knowledge transfer and facilitates faster learning compared to baselines of fixed sequence task learning and isolated task learners with no knowledge transfer.","PeriodicalId":505134,"journal":{"name":"2023 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"83 8","pages":"1-7"},"PeriodicalIF":0.0000,"publicationDate":"2023-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Model for Cognitively Valid Lifelong Learning\",\"authors\":\"Hanne Say, E. Oztop\",\"doi\":\"10.1109/ROBIO58561.2023.10355028\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In continual learning, usually a sequence of tasks are given to a learning agent and the performance of the agent after learning is measured in terms of resistance to catastrophic forgetting, efficacy of knowledge transfer and overall performance on the individual tasks. On the other hand, in multi-task learning, the system is designed to simultaneously acquire knowledge in multiple tasks, often through offline batch learning. A more cognitively valid scenario for lifelong robot learning would be to have a robotic agent to autonomously decide which task to engage and disengage while leveraging many-to-many knowledge transfer ability among tasks during online learning. In this study, we propose a novel lifelong robot learning architecture to fulfill the aforementioned desiderata, and show its validity in an environment where a robot learns the effects of its actions in different task settings. To realize the proposed model, we adopt learning progress measure for task selection, and have the tasks learn by independent neural networks with special structure that allows access to the neural layers of the non-selected tasks. The experiments conducted with a simulated robot arm in an object interaction scenario show that the proposed architecture yields better knowledge transfer and facilitates faster learning compared to baselines of fixed sequence task learning and isolated task learners with no knowledge transfer.\",\"PeriodicalId\":505134,\"journal\":{\"name\":\"2023 IEEE International Conference on Robotics and Biomimetics (ROBIO)\",\"volume\":\"83 8\",\"pages\":\"1-7\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-12-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 IEEE International Conference on Robotics and Biomimetics (ROBIO)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ROBIO58561.2023.10355028\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE International Conference on Robotics and Biomimetics (ROBIO)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ROBIO58561.2023.10355028","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
In continual learning, usually a sequence of tasks are given to a learning agent and the performance of the agent after learning is measured in terms of resistance to catastrophic forgetting, efficacy of knowledge transfer and overall performance on the individual tasks. On the other hand, in multi-task learning, the system is designed to simultaneously acquire knowledge in multiple tasks, often through offline batch learning. A more cognitively valid scenario for lifelong robot learning would be to have a robotic agent to autonomously decide which task to engage and disengage while leveraging many-to-many knowledge transfer ability among tasks during online learning. In this study, we propose a novel lifelong robot learning architecture to fulfill the aforementioned desiderata, and show its validity in an environment where a robot learns the effects of its actions in different task settings. To realize the proposed model, we adopt learning progress measure for task selection, and have the tasks learn by independent neural networks with special structure that allows access to the neural layers of the non-selected tasks. The experiments conducted with a simulated robot arm in an object interaction scenario show that the proposed architecture yields better knowledge transfer and facilitates faster learning compared to baselines of fixed sequence task learning and isolated task learners with no knowledge transfer.