Recommendation systems compliant with legal and editorial policies: the BBC+ app journey

Maria Panteli
{"title":"Recommendation systems compliant with legal and editorial policies: the BBC+ app journey","authors":"Maria Panteli","doi":"10.1145/3298689.3346961","DOIUrl":null,"url":null,"abstract":"The BBC produces thousands of pieces of content every day and numerous BBC products deliver this content to millions of users. For many years the content has been manually curated (this is evident in the selection of stories on the front page of the BBC News website and app for example). To support content creation and curation, a set of editorial guidelines have been developed to build quality and trust in the BBC. As personalisation becomes more important for audience engagement, we have been exploring how algorithmically-driven recommendations could be integrated in our products. In this talk we describe how we developed recommendation systems for the BBC+ app that comply with legal and editorial policies and promote the values of the organisation. We also discuss the challenges we face moving forward, extending the use of recommendation systems for a public service media organisation like the BBC. The BBC+ app is the first product to host in-house recommendations in a fully algorithmically-driven application. The app surfaces short video clips and is targeted at younger audiences. The first challenge we dealt with was content metadata. Content metadata are created for different purposes and managed by different teams across the organisation making it difficult to have reliable and consistent information. Metadata enrichment strategies have been applied to identify content that is considered to be editorially sensitive, such as political content, current legal cases, archived news, commercial content, and content unsuitable for an under 16 audience. Metadata enrichment is also applied to identify content that due care has not been taken such as poor titles, and spelling and grammar mistakes. The first versions of recommendation algorithms exclude all editorially risky content from the recommendations, the most serious of which is avoiding contempt of court. In other cases we exclude content that could undermine our quality and trustworthiness. The General Data Protection Regulation (GDPR) that recently came into effect had strong implications on the design of our system architecture, the choice of the recommendation models, and the implementation of specific product features. For example, the user should be able to delete their data or switch off personalisation at any time. Our system architecture should allow us to trace down and delete all data from that user and switch to non-personalised content. The recommendations should also be explainable and this led us to sometimes choosing a simpler model so that it is possible to more easily explain why a user was recommended a particular type of content. Specific product features were also added to enhance transparency and explainability. For example, the user could view their history of watched items, delete any item, and get an explanation of why a piece of content was recommended to them. At the BBC we aim to not only entertain our audiences but also to inform and educate. These BBC values are also reflected in our evaluation strategies and metrics. While we aim to increase audience engagement we are also responsible for providing recent and diverse content that meets the needs of all our audiences. Accuracy metrics such as Hit Rate and Normalized Discounted Cumulative Gain (NDCG) can give a good estimate of the predictive performance of the model. However, recency and diversity metrics have sometimes more weight in our products, especially in applications delivering news content. What is more, qualitative evaluation is very important before releasing any new model into production. We work closely with editorial teams who provide feedback on the quality of the recommendations and flag content not adhering to the BBC's values or the legal and editorial policies. The development of the BBC+ app has been a great journey. We learned a lot about our content metadata, the implications of GDPR in our system, and our evaluation strategies. We created a minimum viable product that is compliant with legal and editorial policies. However, a lot needs to be done to ensure the recommendations meet the quality standards of the BBC. While excluding editorially sensitive content has limited the risk of contempt of court, algorithmic fairness and impartiality still need to be addressed. We encourage the community to look more into these topics and help us create the way forward towards applications with responsible machine learning.","PeriodicalId":215384,"journal":{"name":"Proceedings of the 13th ACM Conference on Recommender Systems","volume":"74 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 13th ACM Conference on Recommender Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3298689.3346961","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The BBC produces thousands of pieces of content every day and numerous BBC products deliver this content to millions of users. For many years the content has been manually curated (this is evident in the selection of stories on the front page of the BBC News website and app for example). To support content creation and curation, a set of editorial guidelines have been developed to build quality and trust in the BBC. As personalisation becomes more important for audience engagement, we have been exploring how algorithmically-driven recommendations could be integrated in our products. In this talk we describe how we developed recommendation systems for the BBC+ app that comply with legal and editorial policies and promote the values of the organisation. We also discuss the challenges we face moving forward, extending the use of recommendation systems for a public service media organisation like the BBC. The BBC+ app is the first product to host in-house recommendations in a fully algorithmically-driven application. The app surfaces short video clips and is targeted at younger audiences. The first challenge we dealt with was content metadata. Content metadata are created for different purposes and managed by different teams across the organisation making it difficult to have reliable and consistent information. Metadata enrichment strategies have been applied to identify content that is considered to be editorially sensitive, such as political content, current legal cases, archived news, commercial content, and content unsuitable for an under 16 audience. Metadata enrichment is also applied to identify content that due care has not been taken such as poor titles, and spelling and grammar mistakes. The first versions of recommendation algorithms exclude all editorially risky content from the recommendations, the most serious of which is avoiding contempt of court. In other cases we exclude content that could undermine our quality and trustworthiness. The General Data Protection Regulation (GDPR) that recently came into effect had strong implications on the design of our system architecture, the choice of the recommendation models, and the implementation of specific product features. For example, the user should be able to delete their data or switch off personalisation at any time. Our system architecture should allow us to trace down and delete all data from that user and switch to non-personalised content. The recommendations should also be explainable and this led us to sometimes choosing a simpler model so that it is possible to more easily explain why a user was recommended a particular type of content. Specific product features were also added to enhance transparency and explainability. For example, the user could view their history of watched items, delete any item, and get an explanation of why a piece of content was recommended to them. At the BBC we aim to not only entertain our audiences but also to inform and educate. These BBC values are also reflected in our evaluation strategies and metrics. While we aim to increase audience engagement we are also responsible for providing recent and diverse content that meets the needs of all our audiences. Accuracy metrics such as Hit Rate and Normalized Discounted Cumulative Gain (NDCG) can give a good estimate of the predictive performance of the model. However, recency and diversity metrics have sometimes more weight in our products, especially in applications delivering news content. What is more, qualitative evaluation is very important before releasing any new model into production. We work closely with editorial teams who provide feedback on the quality of the recommendations and flag content not adhering to the BBC's values or the legal and editorial policies. The development of the BBC+ app has been a great journey. We learned a lot about our content metadata, the implications of GDPR in our system, and our evaluation strategies. We created a minimum viable product that is compliant with legal and editorial policies. However, a lot needs to be done to ensure the recommendations meet the quality standards of the BBC. While excluding editorially sensitive content has limited the risk of contempt of court, algorithmic fairness and impartiality still need to be addressed. We encourage the community to look more into these topics and help us create the way forward towards applications with responsible machine learning.
符合法律和编辑政策的推荐系统:BBC+应用程序之旅
BBC每天生产成千上万的内容,无数的BBC产品将这些内容提供给数百万用户。多年来,新闻内容都是人工管理的(例如,BBC新闻网站和应用程序首页的故事选择就是很明显的例子)。为了支持内容创作和策划,我们制定了一套编辑指南,以建立BBC的质量和信任。随着个性化对用户参与变得越来越重要,我们一直在探索如何将算法驱动的推荐整合到我们的产品中。在这次演讲中,我们描述了我们如何为BBC+应用程序开发推荐系统,该系统符合法律和编辑政策,并促进了组织的价值观。我们还讨论了我们前进中面临的挑战,将推荐系统的使用扩展到像BBC这样的公共服务媒体组织。BBC+应用程序是第一个在完全由算法驱动的应用程序中托管内部推荐的产品。这款应用展示了短视频片段,目标受众是更年轻的用户。我们处理的第一个挑战是内容元数据。内容元数据是为不同的目的而创建的,并由组织中的不同团队管理,因此很难获得可靠和一致的信息。元数据浓缩策略已被用于识别被认为是编辑敏感的内容,如政治内容、当前法律案件、存档新闻、商业内容和不适合16岁以下观众的内容。元数据丰富还用于识别未得到适当注意的内容,例如糟糕的标题、拼写和语法错误。第一版推荐算法将所有具有编辑风险的内容从推荐中排除,其中最严重的是避免藐视法庭。在其他情况下,我们会排除可能损害我们的质量和可信度的内容。最近生效的通用数据保护条例(GDPR)对我们系统架构的设计、推荐模型的选择和特定产品功能的实现产生了强烈的影响。例如,用户应该能够随时删除他们的数据或关闭个性化设置。我们的系统架构应该允许我们追踪和删除该用户的所有数据,并切换到非个性化的内容。建议也应该是可解释的,这导致我们有时选择一个更简单的模型,以便更容易地解释为什么向用户推荐特定类型的内容。还添加了特定的产品功能,以提高透明度和可解释性。例如,用户可以查看他们观看的项目的历史记录,删除任何项目,并获得为什么向他们推荐一段内容的解释。在BBC,我们的目标不仅是娱乐观众,还要提供信息和教育。这些BBC价值观也反映在我们的评估策略和指标中。虽然我们的目标是增加观众的参与度,但我们也负责提供满足所有观众需求的最新和多样化的内容。命中率和归一化贴现累积增益(NDCG)等精度指标可以很好地估计模型的预测性能。然而,近代性和多样性指标有时在我们的产品中更有分量,特别是在提供新闻内容的应用程序中。此外,在将任何新模型投入生产之前,进行定性评估是非常重要的。我们与编辑团队密切合作,他们对推荐的质量提供反馈,并标记不符合BBC价值观或法律和编辑政策的内容。BBC+应用程序的开发是一段伟大的旅程。我们从内容元数据、GDPR对我们系统的影响以及评估策略等方面学到了很多东西。我们创造了一个符合法律和编辑政策的最小可行产品。然而,要确保这些建议符合BBC的质量标准,还有很多工作要做。虽然排除编辑敏感内容限制了藐视法庭的风险,但算法的公平性和公正性仍然需要解决。我们鼓励社区对这些主题进行更多的研究,并帮助我们为负责任的机器学习应用程序创造前进的道路。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信