Arnau Ramisa, Rene Vidal, Yashar Deldjoo, Zhankui He, Julian McAuley, Anton Korikov, Scott Sanner, Mahesh Sathiamoorthy, Atoosa Kasrizadeh, Silvia Milano, Francesco Ricci
{"title":"推荐系统中的多模式生成模型","authors":"Arnau Ramisa, Rene Vidal, Yashar Deldjoo, Zhankui He, Julian McAuley, Anton Korikov, Scott Sanner, Mahesh Sathiamoorthy, Atoosa Kasrizadeh, Silvia Milano, Francesco Ricci","doi":"arxiv-2409.10993","DOIUrl":null,"url":null,"abstract":"Many recommendation systems limit user inputs to text strings or behavior\nsignals such as clicks and purchases, and system outputs to a list of products\nsorted by relevance. With the advent of generative AI, users have come to\nexpect richer levels of interactions. In visual search, for example, a user may\nprovide a picture of their desired product along with a natural language\nmodification of the content of the picture (e.g., a dress like the one shown in\nthe picture but in red color). Moreover, users may want to better understand\nthe recommendations they receive by visualizing how the product fits their use\ncase, e.g., with a representation of how a garment might look on them, or how a\nfurniture item might look in their room. Such advanced levels of interaction\nrequire recommendation systems that are able to discover both shared and\ncomplementary information about the product across modalities, and visualize\nthe product in a realistic and informative way. However, existing systems often\ntreat multiple modalities independently: text search is usually done by\ncomparing the user query to product titles and descriptions, while visual\nsearch is typically done by comparing an image provided by the customer to\nproduct images. We argue that future recommendation systems will benefit from a\nmulti-modal understanding of the products that leverages the rich information\nretailers have about both customers and products to come up with the best\nrecommendations. In this chapter we review recommendation systems that use\nmultiple data modalities simultaneously.","PeriodicalId":501281,"journal":{"name":"arXiv - CS - Information Retrieval","volume":"11 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-modal Generative Models in Recommendation System\",\"authors\":\"Arnau Ramisa, Rene Vidal, Yashar Deldjoo, Zhankui He, Julian McAuley, Anton Korikov, Scott Sanner, Mahesh Sathiamoorthy, Atoosa Kasrizadeh, Silvia Milano, Francesco Ricci\",\"doi\":\"arxiv-2409.10993\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Many recommendation systems limit user inputs to text strings or behavior\\nsignals such as clicks and purchases, and system outputs to a list of products\\nsorted by relevance. With the advent of generative AI, users have come to\\nexpect richer levels of interactions. In visual search, for example, a user may\\nprovide a picture of their desired product along with a natural language\\nmodification of the content of the picture (e.g., a dress like the one shown in\\nthe picture but in red color). Moreover, users may want to better understand\\nthe recommendations they receive by visualizing how the product fits their use\\ncase, e.g., with a representation of how a garment might look on them, or how a\\nfurniture item might look in their room. Such advanced levels of interaction\\nrequire recommendation systems that are able to discover both shared and\\ncomplementary information about the product across modalities, and visualize\\nthe product in a realistic and informative way. However, existing systems often\\ntreat multiple modalities independently: text search is usually done by\\ncomparing the user query to product titles and descriptions, while visual\\nsearch is typically done by comparing an image provided by the customer to\\nproduct images. We argue that future recommendation systems will benefit from a\\nmulti-modal understanding of the products that leverages the rich information\\nretailers have about both customers and products to come up with the best\\nrecommendations. In this chapter we review recommendation systems that use\\nmultiple data modalities simultaneously.\",\"PeriodicalId\":501281,\"journal\":{\"name\":\"arXiv - CS - Information Retrieval\",\"volume\":\"11 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Information Retrieval\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.10993\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Information Retrieval","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10993","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multi-modal Generative Models in Recommendation System
Many recommendation systems limit user inputs to text strings or behavior
signals such as clicks and purchases, and system outputs to a list of products
sorted by relevance. With the advent of generative AI, users have come to
expect richer levels of interactions. In visual search, for example, a user may
provide a picture of their desired product along with a natural language
modification of the content of the picture (e.g., a dress like the one shown in
the picture but in red color). Moreover, users may want to better understand
the recommendations they receive by visualizing how the product fits their use
case, e.g., with a representation of how a garment might look on them, or how a
furniture item might look in their room. Such advanced levels of interaction
require recommendation systems that are able to discover both shared and
complementary information about the product across modalities, and visualize
the product in a realistic and informative way. However, existing systems often
treat multiple modalities independently: text search is usually done by
comparing the user query to product titles and descriptions, while visual
search is typically done by comparing an image provided by the customer to
product images. We argue that future recommendation systems will benefit from a
multi-modal understanding of the products that leverages the rich information
retailers have about both customers and products to come up with the best
recommendations. In this chapter we review recommendation systems that use
multiple data modalities simultaneously.