Doubly constrained offline reinforcement learning for learning path recommendation

IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Yue Yun, Huan Dai, Rui An, Yupei Zhang, Xuequn Shang
{"title":"Doubly constrained offline reinforcement learning for learning path recommendation","authors":"Yue Yun,&nbsp;Huan Dai,&nbsp;Rui An,&nbsp;Yupei Zhang,&nbsp;Xuequn Shang","doi":"10.1016/j.knosys.2023.111242","DOIUrl":null,"url":null,"abstract":"<div><p>Learning path recommendation refers to the application of interactive recommendation systems in the field of education, aimed at optimizing learning outcomes<span> while minimizing the workload of learners, teachers, and curriculum designers. Reinforcement Learning<span> (RL) has proven effective in capturing and modeling the complex interactions among course activities, learner behaviors, and educational outcomes. Therefore, combining the two approaches presents endless possibilities for personalized education through the use of interactive recommendation systems in the education domain. However, traditional RL algorithms require extensive interaction with the environment during the training phase. Using unverified recommendation logic in interactions with actual students can give rise to unmanageable problems and hinder effective performance in an educational setting. This is because extrapolation introduces substantial evaluation errors that result in recommendations deviating significantly from the actual educational requirements. To address this limitation, we propose a novel method of offline reinforcement learning called Doubly Constrained deep Q-learning Network (DCQN). This method utilizes two generative models to fit existing student historical interaction data, which in turn, constrains the original policy network to generate new actions based on past interactions, avoiding the occurrence of overestimated actions and reducing extrapolation errors. Empirical results on demonstrate that this approach performs better than existing techniques across D4RL, i.e., datasets for deep data-driven reinforcement learning and real educational datasets.</span></span></p></div>","PeriodicalId":49939,"journal":{"name":"Knowledge-Based Systems","volume":"284 ","pages":"Article 111242"},"PeriodicalIF":7.6000,"publicationDate":"2023-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Knowledge-Based Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0950705123009917","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Learning path recommendation refers to the application of interactive recommendation systems in the field of education, aimed at optimizing learning outcomes while minimizing the workload of learners, teachers, and curriculum designers. Reinforcement Learning (RL) has proven effective in capturing and modeling the complex interactions among course activities, learner behaviors, and educational outcomes. Therefore, combining the two approaches presents endless possibilities for personalized education through the use of interactive recommendation systems in the education domain. However, traditional RL algorithms require extensive interaction with the environment during the training phase. Using unverified recommendation logic in interactions with actual students can give rise to unmanageable problems and hinder effective performance in an educational setting. This is because extrapolation introduces substantial evaluation errors that result in recommendations deviating significantly from the actual educational requirements. To address this limitation, we propose a novel method of offline reinforcement learning called Doubly Constrained deep Q-learning Network (DCQN). This method utilizes two generative models to fit existing student historical interaction data, which in turn, constrains the original policy network to generate new actions based on past interactions, avoiding the occurrence of overestimated actions and reducing extrapolation errors. Empirical results on demonstrate that this approach performs better than existing techniques across D4RL, i.e., datasets for deep data-driven reinforcement learning and real educational datasets.

用于学习路径推荐的双约束离线强化学习
学习路径推荐是指交互式推荐系统在教育领域的应用,旨在优化学习成果,同时最大限度地减少学习者、教师和课程设计者的工作量。强化学习(RL)在捕获和建模课程活动、学习者行为和教育成果之间的复杂相互作用方面已被证明是有效的。因此,结合这两种方法,通过在教育领域使用交互式推荐系统,为个性化教育提供了无限的可能性。然而,传统的强化学习算法需要在训练阶段与环境进行广泛的交互。在与实际学生的交互中使用未经验证的推荐逻辑可能会产生难以管理的问题,并阻碍教育环境中的有效表现。这是因为外推法引入了大量的评估错误,导致建议与实际教育要求严重偏离。为了解决这一限制,我们提出了一种新的离线强化学习方法,称为双重约束深度q -学习网络(DCQN)。该方法利用两个生成模型拟合现有的学生历史交互数据,进而约束原始政策网络根据过去的交互生成新的动作,避免了高估动作的发生,减少了外推误差。实证结果表明,该方法在D4RL(即深度数据驱动强化学习的数据集和真实教育数据集)中的表现优于现有技术。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Knowledge-Based Systems
Knowledge-Based Systems 工程技术-计算机:人工智能
CiteScore
14.80
自引率
12.50%
发文量
1245
审稿时长
7.8 months
期刊介绍: Knowledge-Based Systems, an international and interdisciplinary journal in artificial intelligence, publishes original, innovative, and creative research results in the field. It focuses on knowledge-based and other artificial intelligence techniques-based systems. The journal aims to support human prediction and decision-making through data science and computation techniques, provide a balanced coverage of theory and practical study, and encourage the development and implementation of knowledge-based intelligence models, methods, systems, and software tools. Applications in business, government, education, engineering, and healthcare are emphasized.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信