Ling Huang , Zhe-Yuan Li , Xiao-Dong Huang , Yuefang Gao , Chang-Dong Wang , Philip S. Yu
{"title":"Autoencoder-based contrastive learning for next basket recommendation","authors":"Ling Huang , Zhe-Yuan Li , Xiao-Dong Huang , Yuefang Gao , Chang-Dong Wang , Philip S. Yu","doi":"10.1016/j.neunet.2025.108166","DOIUrl":null,"url":null,"abstract":"<div><div>Next Basket Recommendation (NBR) aims to predict the items in the next basket a user will interact with, based on the user’s basket interaction history. However, data sparsity has been a significant challenge in this area. Contrastive Learning (CL) leverages data augmentation and constructs contrastive losses to enhance the embeddings quality, thus effectively addressing the issue of data sparsity. However, the existing methods rely on adding information to basket embedding or segmenting baskets for contrastive learning, which tend to disrupt the original embedding and have limited applicability in the NBR scenarios with diverse data characteristics. To address the above problems, we propose a novel model called Autoencoder-based Contrastive learning for Next Basket Recommendation (AC-NBR). The proposed method mainly consists of three modules, namely AE-based Basket Augmentation, AE-based Contrastive Learning, and Next-Basket Predictor. In the first module, two different basket augmentation methods are designed to provide sufficient and diverse positive pairs for CL. Specifically, we leverage an encoder-decoder structure with appropriate Gaussian noise to extract key features. This process not only helps mitigate noise interference but also improves the robustness of the embedding representation. In addition, the mean and standard deviation of the embedding representation space are learned separately. Then, Gaussian sampling is performed and the sampled latent representation is reconstructed through the decoder to achieve basket augmentation. This approach preserves core information while enhancing the embedding’s diversity and adaptability. In the second module, based on the two basket augmentations and the initial basket embeddings, three sets of positive pairs are constructed for CL. In the third module, we first encode the optimized basket sequence through a Gated Recurrent Unit (GRU) and then employ two Multi-Layer Perceptrons (MLPs) to predict the items likely to be contained in the next basket, thereby obtaining the final prediction results. The effectiveness of AC-NBR is confirmed through comprehensive experiments on three real-world datasets.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"194 ","pages":"Article 108166"},"PeriodicalIF":6.3000,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0893608025010469","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Next Basket Recommendation (NBR) aims to predict the items in the next basket a user will interact with, based on the user’s basket interaction history. However, data sparsity has been a significant challenge in this area. Contrastive Learning (CL) leverages data augmentation and constructs contrastive losses to enhance the embeddings quality, thus effectively addressing the issue of data sparsity. However, the existing methods rely on adding information to basket embedding or segmenting baskets for contrastive learning, which tend to disrupt the original embedding and have limited applicability in the NBR scenarios with diverse data characteristics. To address the above problems, we propose a novel model called Autoencoder-based Contrastive learning for Next Basket Recommendation (AC-NBR). The proposed method mainly consists of three modules, namely AE-based Basket Augmentation, AE-based Contrastive Learning, and Next-Basket Predictor. In the first module, two different basket augmentation methods are designed to provide sufficient and diverse positive pairs for CL. Specifically, we leverage an encoder-decoder structure with appropriate Gaussian noise to extract key features. This process not only helps mitigate noise interference but also improves the robustness of the embedding representation. In addition, the mean and standard deviation of the embedding representation space are learned separately. Then, Gaussian sampling is performed and the sampled latent representation is reconstructed through the decoder to achieve basket augmentation. This approach preserves core information while enhancing the embedding’s diversity and adaptability. In the second module, based on the two basket augmentations and the initial basket embeddings, three sets of positive pairs are constructed for CL. In the third module, we first encode the optimized basket sequence through a Gated Recurrent Unit (GRU) and then employ two Multi-Layer Perceptrons (MLPs) to predict the items likely to be contained in the next basket, thereby obtaining the final prediction results. The effectiveness of AC-NBR is confirmed through comprehensive experiments on three real-world datasets.
Next Basket Recommendation (NBR)旨在根据用户的购物篮交互历史,预测用户将与之交互的下一个购物篮中的商品。然而,数据稀疏性一直是该领域的一个重大挑战。对比学习(CL)利用数据增强和构造对比损失来提高嵌入质量,从而有效地解决了数据稀疏性问题。然而,现有的方法依赖于将信息添加到篮嵌入或分割篮中进行对比学习,这往往会破坏原有的嵌入,并且在具有不同数据特征的NBR场景中的适用性有限。为了解决上述问题,我们提出了一种新的基于自编码器的下一篮推荐对比学习模型(AC-NBR)。该方法主要包括基于ae的篮框增强、基于ae的对比学习和下一个篮框预测三个模块。在第一个模块中,设计了两种不同的篮子扩增方法,为CL提供充足和多样化的正对。具体来说,我们利用具有适当高斯噪声的编码器-解码器结构来提取关键特征。这一过程不仅有助于减轻噪声干扰,而且提高了嵌入表示的鲁棒性。此外,还分别学习了嵌入表示空间的均值和标准差。然后进行高斯采样,通过解码器重构采样后的潜在表示,实现篮增强。该方法在保留核心信息的同时,增强了嵌入的多样性和适应性。在第二个模块中,基于两个篮增广和初始篮嵌入,构造了CL的三组正对。在第三个模块中,我们首先通过门控循环单元(GRU)对优化后的篮子序列进行编码,然后使用两个多层感知器(mlp)预测下一个篮子中可能包含的物品,从而获得最终的预测结果。通过三个真实数据集的综合实验,验证了AC-NBR的有效性。
期刊介绍:
Neural Networks is a platform that aims to foster an international community of scholars and practitioners interested in neural networks, deep learning, and other approaches to artificial intelligence and machine learning. Our journal invites submissions covering various aspects of neural networks research, from computational neuroscience and cognitive modeling to mathematical analyses and engineering applications. By providing a forum for interdisciplinary discussions between biology and technology, we aim to encourage the development of biologically-inspired artificial intelligence.