Proximal Policy Optimization for Explainable Recommended Systems

Qian Feng, Geyang Xiao, Yuan Liang, Huifeng Zhang, Linlin Yan, Xiaoyu Yi
{"title":"Proximal Policy Optimization for Explainable Recommended Systems","authors":"Qian Feng, Geyang Xiao, Yuan Liang, Huifeng Zhang, Linlin Yan, Xiaoyu Yi","doi":"10.1109/DOCS55193.2022.9967709","DOIUrl":null,"url":null,"abstract":"In this paper, in order to intuitively improve the interpretability of the recommendation, we make full use of the knowledge graph and provide an ‘explicit’ recommendation path for the recommended result. Specifically, we embeds the entities and relations in the knowledge graph according to consumers’ behavior. Then a knowledge reasoning method based on reinforcement learning algorithm is proposed, which aims to find a reasonable path after two entities in the knowledge graph are given. The main contributions of this paper are as follows: The current State-of-the-art method in reinforcement learning, PPO is used for improvement. The soft-reward function after adjustment behaves better for the current problem while training. And a Q-value based path exploration method is innovated for testing. With extensive experiments on several real-world datasets provided by Amazon, our method behaves better for the current problem compared with the baseline.","PeriodicalId":348545,"journal":{"name":"2022 4th International Conference on Data-driven Optimization of Complex Systems (DOCS)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 4th International Conference on Data-driven Optimization of Complex Systems (DOCS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DOCS55193.2022.9967709","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In this paper, in order to intuitively improve the interpretability of the recommendation, we make full use of the knowledge graph and provide an ‘explicit’ recommendation path for the recommended result. Specifically, we embeds the entities and relations in the knowledge graph according to consumers’ behavior. Then a knowledge reasoning method based on reinforcement learning algorithm is proposed, which aims to find a reasonable path after two entities in the knowledge graph are given. The main contributions of this paper are as follows: The current State-of-the-art method in reinforcement learning, PPO is used for improvement. The soft-reward function after adjustment behaves better for the current problem while training. And a Q-value based path exploration method is innovated for testing. With extensive experiments on several real-world datasets provided by Amazon, our method behaves better for the current problem compared with the baseline.
可解释推荐系统的近端策略优化
在本文中,为了直观地提高推荐的可解释性,我们充分利用了知识图,为推荐结果提供了一条“显式”的推荐路径。具体来说,我们根据消费者的行为将实体和关系嵌入到知识图中。然后提出了一种基于强化学习算法的知识推理方法,其目的是在给定知识图中的两个实体后,找到一条合理的路径。本文的主要贡献如下:使用当前最先进的强化学习方法,PPO进行改进。在训练过程中,调整后的软奖励函数对当前问题的表现更好。创新了一种基于q值的路径探索方法。在Amazon提供的几个真实数据集上进行了大量的实验,与基线相比,我们的方法对当前问题的表现更好。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信