{"title":"ACM RecSys CrowdRec 2015研讨会综述:推荐系统的众包和人工计算","authors":"M. Larson, D. Tikk, R. Turrin","doi":"10.1145/2792838.2798719","DOIUrl":null,"url":null,"abstract":"CrowdRec 2015 provides the recommender system community with a forum at which to discuss crowdsourcing and human computation. Systems that explicitly collect information from human annotators to improve recommendations are becoming more widespread. At this year's workshop, we highlight incentivization and the issue of avoiding bias. We take a special look at how recommender systems can influence collective behavior, and the contribution that the crowd can make to recommender system evaluation.","PeriodicalId":325637,"journal":{"name":"Proceedings of the 9th ACM Conference on Recommender Systems","volume":"74 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Overview of ACM RecSys CrowdRec 2015 Workshop: Crowdsourcing and Human Computation for Recommender Systems\",\"authors\":\"M. Larson, D. Tikk, R. Turrin\",\"doi\":\"10.1145/2792838.2798719\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"CrowdRec 2015 provides the recommender system community with a forum at which to discuss crowdsourcing and human computation. Systems that explicitly collect information from human annotators to improve recommendations are becoming more widespread. At this year's workshop, we highlight incentivization and the issue of avoiding bias. We take a special look at how recommender systems can influence collective behavior, and the contribution that the crowd can make to recommender system evaluation.\",\"PeriodicalId\":325637,\"journal\":{\"name\":\"Proceedings of the 9th ACM Conference on Recommender Systems\",\"volume\":\"74 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2015-09-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 9th ACM Conference on Recommender Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/2792838.2798719\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 9th ACM Conference on Recommender Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/2792838.2798719","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Overview of ACM RecSys CrowdRec 2015 Workshop: Crowdsourcing and Human Computation for Recommender Systems
CrowdRec 2015 provides the recommender system community with a forum at which to discuss crowdsourcing and human computation. Systems that explicitly collect information from human annotators to improve recommendations are becoming more widespread. At this year's workshop, we highlight incentivization and the issue of avoiding bias. We take a special look at how recommender systems can influence collective behavior, and the contribution that the crowd can make to recommender system evaluation.