Ask the GRU: Multi-task Learning for Deep Text Recommendations

Trapit Bansal, David Belanger, A. McCallum
{"title":"Ask the GRU: Multi-task Learning for Deep Text Recommendations","authors":"Trapit Bansal, David Belanger, A. McCallum","doi":"10.1145/2959100.2959180","DOIUrl":null,"url":null,"abstract":"In a variety of application domains the content to be recommended to users is associated with text. This includes research papers, movies with associated plot summaries, news articles, blog posts, etc. Recommendation approaches based on latent factor models can be extended naturally to leverage text by employing an explicit mapping from text to factors. This enables recommendations for new, unseen content, and may generalize better, since the factors for all items are produced by a compactly-parametrized model. Previous work has used topic models or averages of word embeddings for this mapping. In this paper we present a method leveraging deep recurrent neural networks to encode the text sequence into a latent vector, specifically gated recurrent units (GRUs) trained end-to-end on the collaborative filtering task. For the task of scientific paper recommendation, this yields models with significantly higher accuracy. In cold-start scenarios, we beat the previous state-of-the-art, all of which ignore word order. Performance is further improved by multi-task learning, where the text encoder network is trained for a combination of content recommendation and item metadata prediction. This regularizes the collaborative filtering model, ameliorating the problem of sparsity of the observed rating matrix.","PeriodicalId":315651,"journal":{"name":"Proceedings of the 10th ACM Conference on Recommender Systems","volume":"63 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2016-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"296","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 10th ACM Conference on Recommender Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/2959100.2959180","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 296

Abstract

In a variety of application domains the content to be recommended to users is associated with text. This includes research papers, movies with associated plot summaries, news articles, blog posts, etc. Recommendation approaches based on latent factor models can be extended naturally to leverage text by employing an explicit mapping from text to factors. This enables recommendations for new, unseen content, and may generalize better, since the factors for all items are produced by a compactly-parametrized model. Previous work has used topic models or averages of word embeddings for this mapping. In this paper we present a method leveraging deep recurrent neural networks to encode the text sequence into a latent vector, specifically gated recurrent units (GRUs) trained end-to-end on the collaborative filtering task. For the task of scientific paper recommendation, this yields models with significantly higher accuracy. In cold-start scenarios, we beat the previous state-of-the-art, all of which ignore word order. Performance is further improved by multi-task learning, where the text encoder network is trained for a combination of content recommendation and item metadata prediction. This regularizes the collaborative filtering model, ameliorating the problem of sparsity of the observed rating matrix.
问GRU:深度文本推荐的多任务学习
在各种应用领域中,向用户推荐的内容与文本相关联。这包括研究论文、带有相关情节摘要的电影、新闻文章、博客文章等。基于潜在因素模型的推荐方法可以通过使用从文本到因素的显式映射自然地扩展到利用文本。这可以推荐新的、未见过的内容,并且可以更好地推广,因为所有项目的因素都是由紧凑参数化模型产生的。以前的工作使用主题模型或词嵌入的平均值来进行这种映射。在本文中,我们提出了一种利用深度递归神经网络将文本序列编码为潜在向量的方法,特别是在协同过滤任务上端到端训练的门控递归单元(gru)。对于科学论文推荐任务,这产生了精度显著提高的模型。在冷启动场景中,我们击败了之前的最先进的技术,所有这些技术都忽略了单词顺序。通过多任务学习进一步提高性能,其中文本编码器网络被训练为内容推荐和项目元数据预测的组合。这使协同过滤模型正则化,改善了观察到的评级矩阵的稀疏性问题。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信