{"title":"通过特征态度在具身系统中更快地学习","authors":"D. Jacob, D. Polani, Chrystopher L. Nehaniv","doi":"10.1109/CIRA.2005.1554338","DOIUrl":null,"url":null,"abstract":"Classical reinforcement learning is a general learning paradigm with wide applicability in many problem domains. Where embodied agents are concerned, however, it is unable to take advantage of the structured, regular nature of the physical world to maximise learning efficiency. Here, using a model of a three joint robot arm, we show initial learning accelerated by an order of magnitude using simple constraints to produce characteristic attitudes, implemented as part of the learning algorithm. We point out possible parallels with constraints on the movement of natural organisms owing to their detailed mechanical structure. The work forms part of our EMBER framework for reinforcement learning in embodied agents introduced and developed in 2004.","PeriodicalId":162553,"journal":{"name":"2005 International Symposium on Computational Intelligence in Robotics and Automation","volume":"14 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2005-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Faster learning in embodied systems through characteristic attitudes\",\"authors\":\"D. Jacob, D. Polani, Chrystopher L. Nehaniv\",\"doi\":\"10.1109/CIRA.2005.1554338\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Classical reinforcement learning is a general learning paradigm with wide applicability in many problem domains. Where embodied agents are concerned, however, it is unable to take advantage of the structured, regular nature of the physical world to maximise learning efficiency. Here, using a model of a three joint robot arm, we show initial learning accelerated by an order of magnitude using simple constraints to produce characteristic attitudes, implemented as part of the learning algorithm. We point out possible parallels with constraints on the movement of natural organisms owing to their detailed mechanical structure. The work forms part of our EMBER framework for reinforcement learning in embodied agents introduced and developed in 2004.\",\"PeriodicalId\":162553,\"journal\":{\"name\":\"2005 International Symposium on Computational Intelligence in Robotics and Automation\",\"volume\":\"14 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2005-12-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2005 International Symposium on Computational Intelligence in Robotics and Automation\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CIRA.2005.1554338\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2005 International Symposium on Computational Intelligence in Robotics and Automation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CIRA.2005.1554338","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Faster learning in embodied systems through characteristic attitudes
Classical reinforcement learning is a general learning paradigm with wide applicability in many problem domains. Where embodied agents are concerned, however, it is unable to take advantage of the structured, regular nature of the physical world to maximise learning efficiency. Here, using a model of a three joint robot arm, we show initial learning accelerated by an order of magnitude using simple constraints to produce characteristic attitudes, implemented as part of the learning algorithm. We point out possible parallels with constraints on the movement of natural organisms owing to their detailed mechanical structure. The work forms part of our EMBER framework for reinforcement learning in embodied agents introduced and developed in 2004.