{"title":"推荐系统的全空间多门专家混合模型","authors":"Zheng Ye, Jun Ge","doi":"10.1109/WI-IAT55865.2022.00047","DOIUrl":null,"url":null,"abstract":"With the development of e-commerce, both advertisers and platforms pay more and more attention to the effectiveness of ads recommendation. In recent years, deep learning approaches with a mulit-task learning framework have shown to be effective in such recommendation systems. One main goal of these systems is to estimate the post-click conversion rate(CVR) accurately. However, higher click-through rate(CTR) for a product does not always lead to higher conversion rate(CVR) due to many reasons (e.g. lower rating). In addition, the overall performance of the recommendation system may not be optimal, since the usage of multi-task models (the CTR and CVR tasks) is often sensitive to the relationships of the tasks. In this paper, we propose a deep neural model under the Mixture-of-Experts framework (MoE), call ES-MMOE, in which a sub-network is used to promote samples with high CVR. The model can also be trained with the entire space by taking advantage of the Entire Space Multi-task Model (ESMM) model. Extensive experiments on a large-scale dataset gathered from traffic logs of Taobao’s recommender system demonstrate that ES-MMOE outperforms a number of the state-of-the-art models, including ESMM, with a relatively large margin.","PeriodicalId":345445,"journal":{"name":"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Entire Space Multi-gate Mixture-of-Experts Model for Recommender Systems\",\"authors\":\"Zheng Ye, Jun Ge\",\"doi\":\"10.1109/WI-IAT55865.2022.00047\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With the development of e-commerce, both advertisers and platforms pay more and more attention to the effectiveness of ads recommendation. In recent years, deep learning approaches with a mulit-task learning framework have shown to be effective in such recommendation systems. One main goal of these systems is to estimate the post-click conversion rate(CVR) accurately. However, higher click-through rate(CTR) for a product does not always lead to higher conversion rate(CVR) due to many reasons (e.g. lower rating). In addition, the overall performance of the recommendation system may not be optimal, since the usage of multi-task models (the CTR and CVR tasks) is often sensitive to the relationships of the tasks. In this paper, we propose a deep neural model under the Mixture-of-Experts framework (MoE), call ES-MMOE, in which a sub-network is used to promote samples with high CVR. The model can also be trained with the entire space by taking advantage of the Entire Space Multi-task Model (ESMM) model. Extensive experiments on a large-scale dataset gathered from traffic logs of Taobao’s recommender system demonstrate that ES-MMOE outperforms a number of the state-of-the-art models, including ESMM, with a relatively large margin.\",\"PeriodicalId\":345445,\"journal\":{\"name\":\"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)\",\"volume\":\"24 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/WI-IAT55865.2022.00047\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology (WI-IAT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WI-IAT55865.2022.00047","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
An Entire Space Multi-gate Mixture-of-Experts Model for Recommender Systems
With the development of e-commerce, both advertisers and platforms pay more and more attention to the effectiveness of ads recommendation. In recent years, deep learning approaches with a mulit-task learning framework have shown to be effective in such recommendation systems. One main goal of these systems is to estimate the post-click conversion rate(CVR) accurately. However, higher click-through rate(CTR) for a product does not always lead to higher conversion rate(CVR) due to many reasons (e.g. lower rating). In addition, the overall performance of the recommendation system may not be optimal, since the usage of multi-task models (the CTR and CVR tasks) is often sensitive to the relationships of the tasks. In this paper, we propose a deep neural model under the Mixture-of-Experts framework (MoE), call ES-MMOE, in which a sub-network is used to promote samples with high CVR. The model can also be trained with the entire space by taking advantage of the Entire Space Multi-task Model (ESMM) model. Extensive experiments on a large-scale dataset gathered from traffic logs of Taobao’s recommender system demonstrate that ES-MMOE outperforms a number of the state-of-the-art models, including ESMM, with a relatively large margin.