{"title":"分解归纳式程序学习:以类似人类的数据效率学习学术任务","authors":"Daniel Weitekamp","doi":"10.1609/aaaiss.v3i1.31289","DOIUrl":null,"url":null,"abstract":"Human brains have many differently functioning regions which play specialized roles in learning. By contrast, methods for training artificial neural networks, such as reinforcement-learning, typically learn exclusively via a single mechanism: gradient descent. This raises the question: might human learners’ advantage in learning efficiency over deep-learning be attributed to the interplay between multiple specialized mechanisms of learning? In this work we review a series of simulated learner systems which have been built with the aim of modeling human student’s inductive learning as they practice STEM procedural tasks. By comparison to modern deep-learning based methods which train on thousands to millions of examples to acquire passing performance capabilities, these simulated learners match human performance curves---achieving passing levels of performance within about a dozen practice opportunities. We investigate this impressive learning efficiency via an ablation analysis. Beginning with end-to-end reinforcement learning (1-mechanism), we decompose learning systems incrementally to construct the 3-mechanism inductive learning characteristic of prior simulated learners such as Sierra, SimStudent and the Apprentice Learner Architecture. Our analysis shows that learning decomposition plays a significant role in achieving data-efficient learning on par with human learners---a greater role even than simple distinctions between symbolic/subsymbolic learning. Finally we highlight how this breakdown in learning mechanisms can flexibly incorporate diverse forms of natural language and interface grounded instruction, and discuss opportunities for using these flexible learning capabilities in interactive task learning systems that learn directly from a user’s natural instruction.","PeriodicalId":516827,"journal":{"name":"Proceedings of the AAAI Symposium Series","volume":"72 4","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Decomposed Inductive Procedure Learning: Learning Academic Tasks with Human-Like Data Efficiency\",\"authors\":\"Daniel Weitekamp\",\"doi\":\"10.1609/aaaiss.v3i1.31289\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Human brains have many differently functioning regions which play specialized roles in learning. By contrast, methods for training artificial neural networks, such as reinforcement-learning, typically learn exclusively via a single mechanism: gradient descent. This raises the question: might human learners’ advantage in learning efficiency over deep-learning be attributed to the interplay between multiple specialized mechanisms of learning? In this work we review a series of simulated learner systems which have been built with the aim of modeling human student’s inductive learning as they practice STEM procedural tasks. By comparison to modern deep-learning based methods which train on thousands to millions of examples to acquire passing performance capabilities, these simulated learners match human performance curves---achieving passing levels of performance within about a dozen practice opportunities. We investigate this impressive learning efficiency via an ablation analysis. Beginning with end-to-end reinforcement learning (1-mechanism), we decompose learning systems incrementally to construct the 3-mechanism inductive learning characteristic of prior simulated learners such as Sierra, SimStudent and the Apprentice Learner Architecture. Our analysis shows that learning decomposition plays a significant role in achieving data-efficient learning on par with human learners---a greater role even than simple distinctions between symbolic/subsymbolic learning. Finally we highlight how this breakdown in learning mechanisms can flexibly incorporate diverse forms of natural language and interface grounded instruction, and discuss opportunities for using these flexible learning capabilities in interactive task learning systems that learn directly from a user’s natural instruction.\",\"PeriodicalId\":516827,\"journal\":{\"name\":\"Proceedings of the AAAI Symposium Series\",\"volume\":\"72 4\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-05-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the AAAI Symposium Series\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1609/aaaiss.v3i1.31289\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the AAAI Symposium Series","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1609/aaaiss.v3i1.31289","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Decomposed Inductive Procedure Learning: Learning Academic Tasks with Human-Like Data Efficiency
Human brains have many differently functioning regions which play specialized roles in learning. By contrast, methods for training artificial neural networks, such as reinforcement-learning, typically learn exclusively via a single mechanism: gradient descent. This raises the question: might human learners’ advantage in learning efficiency over deep-learning be attributed to the interplay between multiple specialized mechanisms of learning? In this work we review a series of simulated learner systems which have been built with the aim of modeling human student’s inductive learning as they practice STEM procedural tasks. By comparison to modern deep-learning based methods which train on thousands to millions of examples to acquire passing performance capabilities, these simulated learners match human performance curves---achieving passing levels of performance within about a dozen practice opportunities. We investigate this impressive learning efficiency via an ablation analysis. Beginning with end-to-end reinforcement learning (1-mechanism), we decompose learning systems incrementally to construct the 3-mechanism inductive learning characteristic of prior simulated learners such as Sierra, SimStudent and the Apprentice Learner Architecture. Our analysis shows that learning decomposition plays a significant role in achieving data-efficient learning on par with human learners---a greater role even than simple distinctions between symbolic/subsymbolic learning. Finally we highlight how this breakdown in learning mechanisms can flexibly incorporate diverse forms of natural language and interface grounded instruction, and discuss opportunities for using these flexible learning capabilities in interactive task learning systems that learn directly from a user’s natural instruction.