Weikang He, Yunpeng Xiao, Tun Li, Rong Wang, Qian Li
{"title":"通过多源数据融合构建行为意向预测的基础模型","authors":"Weikang He, Yunpeng Xiao, Tun Li, Rong Wang, Qian Li","doi":"10.1016/j.inffus.2025.103799","DOIUrl":null,"url":null,"abstract":"<div><div>In recent years, some studies have attempted to build foundation models to predict user behavioral intentions across various scenarios. Existing methods typically involve two steps: first, using Pre-trained Language Models (PLMs) to convert user historical behaviors into semantic vectors; second, capturing latent user intentions through a behavioral sequence encoder. However, limitations still exist in this pipeline. Firstly, most studies directly fix the parameters of PLMs, making it difficult to optimize the generated semantic vectors for specific application scenarios. Even with fine-tuning PLMs considered, the huge model size brings computational burdens. To address this challenge, we introduce a Low-Rank Adaptation technique, which enhances the quality of semantic vectors by training two lightweight low-rank matrices while keeping the original parameters of PLMs frozen. Additionally, the semantic vectors generated by PLMs often exhibit anisotropy, which weakens the expressive power. To address this, we design a parameter whitening expert and a mixed observation expert module to achieve more isotropic semantic representations. Finally, there are significant differences in user behavior patterns across different domains, and such behavioral conflicts can limit the model’s generalization ability. Therefore, during the pre-training phase, we introduce multi-domain negative sample fusion and design two contrastive learning tasks. By employing a multi-task learning strategy, we enhance the model’s ability to handle multi-domain data. With these improvements, our model achieves stronger adaptability and accuracy in predicting user behavior across different domains.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"127 ","pages":"Article 103799"},"PeriodicalIF":15.5000,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Towards a foundation model for behavioral intention prediction through multi-source data fusion\",\"authors\":\"Weikang He, Yunpeng Xiao, Tun Li, Rong Wang, Qian Li\",\"doi\":\"10.1016/j.inffus.2025.103799\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>In recent years, some studies have attempted to build foundation models to predict user behavioral intentions across various scenarios. Existing methods typically involve two steps: first, using Pre-trained Language Models (PLMs) to convert user historical behaviors into semantic vectors; second, capturing latent user intentions through a behavioral sequence encoder. However, limitations still exist in this pipeline. Firstly, most studies directly fix the parameters of PLMs, making it difficult to optimize the generated semantic vectors for specific application scenarios. Even with fine-tuning PLMs considered, the huge model size brings computational burdens. To address this challenge, we introduce a Low-Rank Adaptation technique, which enhances the quality of semantic vectors by training two lightweight low-rank matrices while keeping the original parameters of PLMs frozen. Additionally, the semantic vectors generated by PLMs often exhibit anisotropy, which weakens the expressive power. To address this, we design a parameter whitening expert and a mixed observation expert module to achieve more isotropic semantic representations. Finally, there are significant differences in user behavior patterns across different domains, and such behavioral conflicts can limit the model’s generalization ability. Therefore, during the pre-training phase, we introduce multi-domain negative sample fusion and design two contrastive learning tasks. By employing a multi-task learning strategy, we enhance the model’s ability to handle multi-domain data. With these improvements, our model achieves stronger adaptability and accuracy in predicting user behavior across different domains.</div></div>\",\"PeriodicalId\":50367,\"journal\":{\"name\":\"Information Fusion\",\"volume\":\"127 \",\"pages\":\"Article 103799\"},\"PeriodicalIF\":15.5000,\"publicationDate\":\"2025-10-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Fusion\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1566253525008619\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525008619","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Towards a foundation model for behavioral intention prediction through multi-source data fusion
In recent years, some studies have attempted to build foundation models to predict user behavioral intentions across various scenarios. Existing methods typically involve two steps: first, using Pre-trained Language Models (PLMs) to convert user historical behaviors into semantic vectors; second, capturing latent user intentions through a behavioral sequence encoder. However, limitations still exist in this pipeline. Firstly, most studies directly fix the parameters of PLMs, making it difficult to optimize the generated semantic vectors for specific application scenarios. Even with fine-tuning PLMs considered, the huge model size brings computational burdens. To address this challenge, we introduce a Low-Rank Adaptation technique, which enhances the quality of semantic vectors by training two lightweight low-rank matrices while keeping the original parameters of PLMs frozen. Additionally, the semantic vectors generated by PLMs often exhibit anisotropy, which weakens the expressive power. To address this, we design a parameter whitening expert and a mixed observation expert module to achieve more isotropic semantic representations. Finally, there are significant differences in user behavior patterns across different domains, and such behavioral conflicts can limit the model’s generalization ability. Therefore, during the pre-training phase, we introduce multi-domain negative sample fusion and design two contrastive learning tasks. By employing a multi-task learning strategy, we enhance the model’s ability to handle multi-domain data. With these improvements, our model achieves stronger adaptability and accuracy in predicting user behavior across different domains.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.