TBformer: Multi-Scale Transformer With Time-Behavior Attention for Multi-Modal Customer Churn Prediction

IF 10.9 2区 计算机科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC
Yushi Li;Yunfei Tao;Ming Zhu;Ziwen Chen;Zhenyu Wen;Bideng Zhu
{"title":"TBformer: Multi-Scale Transformer With Time-Behavior Attention for Multi-Modal Customer Churn Prediction","authors":"Yushi Li;Yunfei Tao;Ming Zhu;Ziwen Chen;Zhenyu Wen;Bideng Zhu","doi":"10.1109/TCE.2025.3563905","DOIUrl":null,"url":null,"abstract":"In highly competitive market of Internet service platforms, identifying and retaining potential churners through customer churn prediction techniques is crucial for maintaining platform vitality. The sequences of interaction behaviors between customers and platforms are closely related to churn prediction results. However, existing methods focus only on capturing the temporal dependencies in dynamic behavior sequences while ignoring the correlations between different behaviors. Moreover, classical methods apply only to static data, while deep learning-based methods focus on dynamic data, neither leveraging the complementary information between static and dynamic data. To address these issues, we propose a multi-modal customer churn prediction model based on Transformer with multi-scale Time-Behavior attention, TBformer, which adaptively fuses static and dynamic data. Time-Behavior module can capture multi-scale temporal dependencies and behavioral correlations in behavioral time series across time and behavior dimensions. We perform behavior-independent multi-scale dynamic feature fusion through bidirectional connection paths. Furthermore, the multi-modal fusion module based on the attention mechanism adaptively controls the fusion weights of static and dynamic features to improve performance. Extensive experiments on two publicly available datasets, KKBox and KDD, and a private dataset, HOF, demonstrate that our TBformer achieves an average AUC of 91.2% (+2.47%), outperforming the state-of-the-art customer churn prediction methods.","PeriodicalId":13208,"journal":{"name":"IEEE Transactions on Consumer Electronics","volume":"71 2","pages":"3192-3203"},"PeriodicalIF":10.9000,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Consumer Electronics","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10976249/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

In highly competitive market of Internet service platforms, identifying and retaining potential churners through customer churn prediction techniques is crucial for maintaining platform vitality. The sequences of interaction behaviors between customers and platforms are closely related to churn prediction results. However, existing methods focus only on capturing the temporal dependencies in dynamic behavior sequences while ignoring the correlations between different behaviors. Moreover, classical methods apply only to static data, while deep learning-based methods focus on dynamic data, neither leveraging the complementary information between static and dynamic data. To address these issues, we propose a multi-modal customer churn prediction model based on Transformer with multi-scale Time-Behavior attention, TBformer, which adaptively fuses static and dynamic data. Time-Behavior module can capture multi-scale temporal dependencies and behavioral correlations in behavioral time series across time and behavior dimensions. We perform behavior-independent multi-scale dynamic feature fusion through bidirectional connection paths. Furthermore, the multi-modal fusion module based on the attention mechanism adaptively controls the fusion weights of static and dynamic features to improve performance. Extensive experiments on two publicly available datasets, KKBox and KDD, and a private dataset, HOF, demonstrate that our TBformer achieves an average AUC of 91.2% (+2.47%), outperforming the state-of-the-art customer churn prediction methods.
TBformer:多模态客户流失预测的多尺度时间行为关注变压器
在竞争激烈的互联网服务平台市场中,通过客户流失预测技术来识别和留住潜在的流失客户对于保持平台的活力至关重要。客户与平台之间的交互行为顺序与流失预测结果密切相关。然而,现有的方法只关注捕获动态行为序列中的时间依赖性,而忽略了不同行为之间的相关性。此外,经典方法仅适用于静态数据,而基于深度学习的方法侧重于动态数据,两者都没有利用静态和动态数据之间的互补信息。为了解决这些问题,我们提出了一种基于多尺度时间行为关注的变压器(Transformer)的多模态客户流失预测模型,该模型自适应融合了静态和动态数据。时间-行为模块可以捕获行为时间序列中跨时间和行为维度的多尺度时间依赖性和行为相关性。通过双向连接路径实现与行为无关的多尺度动态特征融合。此外,基于注意机制的多模态融合模块自适应控制静态和动态特征的融合权重,以提高性能。在两个公开可用的数据集KKBox和KDD以及一个私人数据集HOF上进行的大量实验表明,我们的TBformer实现了91.2%(+2.47%)的平均AUC,优于最先进的客户流失预测方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
7.70
自引率
9.30%
发文量
59
审稿时长
3.3 months
期刊介绍: The main focus for the IEEE Transactions on Consumer Electronics is the engineering and research aspects of the theory, design, construction, manufacture or end use of mass market electronics, systems, software and services for consumers.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信