{"title":"基于多元PID控制器的随机梯度下降潜因子分析的学习误差细化","authors":"Jinli Li;Ye Yuan;Xin Luo","doi":"10.1109/TETCI.2025.3547854","DOIUrl":null,"url":null,"abstract":"In Big Data-based applications, high-dimensional and incomplete (HDI) data are frequently used to represent the complicated interactions among numerous nodes. A stochastic gradient descent (SGD)-based latent factor analysis (LFA) model can process such data efficiently. Unfortunately, a standard SGD algorithm trains a single latent factor relying on the stochastic gradient related to the current learning error only, leading to a slow convergence rate. To break through this bottleneck, this study establishes an SGD-based LFA model as the backbone, and proposes six proportional-integral-derivative (PID)-incorporated LFA models with diversified PID-controllers with the following two-fold ideas: a) refining the instant learning error in stochastic gradient by the principle of six PID-variants, i.e., a standard PID, an integral separated PID, a gearshift integral PID, a dead zone PID, an anti-windup PID, and an incomplete differential PID, to assimilate historical update information into the learning scheme in an efficient way; b) making the hyper-parameters adaptation by utilizing the mechanism of particle swarm optimization for acquiring high practicality. In addition, considering the diversified PID-variants, an effective ensemble is implemented for the six PID-incorporated LFA models. Experimental results on industrial HDI datasets illustrate that in comparison with state-of-the-art models, the proposed models obtain superior computational efficiency while maintaining competitive accuracy in predicting missing data within an HDI matrix. Moreover, their ensemble further improves performance in terms of prediction accuracy.","PeriodicalId":13135,"journal":{"name":"IEEE Transactions on Emerging Topics in Computational Intelligence","volume":"9 5","pages":"3582-3597"},"PeriodicalIF":5.3000,"publicationDate":"2025-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Learning Error Refinement in Stochastic Gradient Descent-Based Latent Factor Analysis via Diversified PID Controllers\",\"authors\":\"Jinli Li;Ye Yuan;Xin Luo\",\"doi\":\"10.1109/TETCI.2025.3547854\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In Big Data-based applications, high-dimensional and incomplete (HDI) data are frequently used to represent the complicated interactions among numerous nodes. A stochastic gradient descent (SGD)-based latent factor analysis (LFA) model can process such data efficiently. Unfortunately, a standard SGD algorithm trains a single latent factor relying on the stochastic gradient related to the current learning error only, leading to a slow convergence rate. To break through this bottleneck, this study establishes an SGD-based LFA model as the backbone, and proposes six proportional-integral-derivative (PID)-incorporated LFA models with diversified PID-controllers with the following two-fold ideas: a) refining the instant learning error in stochastic gradient by the principle of six PID-variants, i.e., a standard PID, an integral separated PID, a gearshift integral PID, a dead zone PID, an anti-windup PID, and an incomplete differential PID, to assimilate historical update information into the learning scheme in an efficient way; b) making the hyper-parameters adaptation by utilizing the mechanism of particle swarm optimization for acquiring high practicality. In addition, considering the diversified PID-variants, an effective ensemble is implemented for the six PID-incorporated LFA models. Experimental results on industrial HDI datasets illustrate that in comparison with state-of-the-art models, the proposed models obtain superior computational efficiency while maintaining competitive accuracy in predicting missing data within an HDI matrix. Moreover, their ensemble further improves performance in terms of prediction accuracy.\",\"PeriodicalId\":13135,\"journal\":{\"name\":\"IEEE Transactions on Emerging Topics in Computational Intelligence\",\"volume\":\"9 5\",\"pages\":\"3582-3597\"},\"PeriodicalIF\":5.3000,\"publicationDate\":\"2025-03-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Emerging Topics in Computational Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10938033/\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Emerging Topics in Computational Intelligence","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10938033/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
摘要
在基于大数据的应用中,经常使用高维不完整数据(high-dimensional and incomplete, HDI)来表示众多节点之间复杂的交互。基于随机梯度下降(SGD)的潜在因素分析(LFA)模型可以有效地处理这些数据。不幸的是,标准的SGD算法只依靠与当前学习误差相关的随机梯度来训练单个潜在因子,导致收敛速度缓慢。为了突破这一瓶颈,本研究建立了一个基于sgd的LFA模型作为主干,并提出了6个包含比例-积分-导数(PID)的LFA模型,采用多种PID控制器,其思路如下:a)利用标准PID、积分分离PID、换挡积分PID、死区PID、反上卷PID、不完全微分PID等6个PID变量的原理,改进随机梯度中的瞬时学习误差,有效地将历史更新信息吸收到学习方案中;B)利用粒子群优化机制进行超参数自适应,以获得较高的实用性。此外,考虑到pid变量的多样性,对6个包含pid的LFA模型进行了有效的集成。工业HDI数据集的实验结果表明,与最先进的模型相比,所提出的模型在预测HDI矩阵中缺失数据时获得了更高的计算效率,同时保持了竞争的准确性。此外,它们的集成进一步提高了预测精度。
Learning Error Refinement in Stochastic Gradient Descent-Based Latent Factor Analysis via Diversified PID Controllers
In Big Data-based applications, high-dimensional and incomplete (HDI) data are frequently used to represent the complicated interactions among numerous nodes. A stochastic gradient descent (SGD)-based latent factor analysis (LFA) model can process such data efficiently. Unfortunately, a standard SGD algorithm trains a single latent factor relying on the stochastic gradient related to the current learning error only, leading to a slow convergence rate. To break through this bottleneck, this study establishes an SGD-based LFA model as the backbone, and proposes six proportional-integral-derivative (PID)-incorporated LFA models with diversified PID-controllers with the following two-fold ideas: a) refining the instant learning error in stochastic gradient by the principle of six PID-variants, i.e., a standard PID, an integral separated PID, a gearshift integral PID, a dead zone PID, an anti-windup PID, and an incomplete differential PID, to assimilate historical update information into the learning scheme in an efficient way; b) making the hyper-parameters adaptation by utilizing the mechanism of particle swarm optimization for acquiring high practicality. In addition, considering the diversified PID-variants, an effective ensemble is implemented for the six PID-incorporated LFA models. Experimental results on industrial HDI datasets illustrate that in comparison with state-of-the-art models, the proposed models obtain superior computational efficiency while maintaining competitive accuracy in predicting missing data within an HDI matrix. Moreover, their ensemble further improves performance in terms of prediction accuracy.
期刊介绍:
The IEEE Transactions on Emerging Topics in Computational Intelligence (TETCI) publishes original articles on emerging aspects of computational intelligence, including theory, applications, and surveys.
TETCI is an electronics only publication. TETCI publishes six issues per year.
Authors are encouraged to submit manuscripts in any emerging topic in computational intelligence, especially nature-inspired computing topics not covered by other IEEE Computational Intelligence Society journals. A few such illustrative examples are glial cell networks, computational neuroscience, Brain Computer Interface, ambient intelligence, non-fuzzy computing with words, artificial life, cultural learning, artificial endocrine networks, social reasoning, artificial hormone networks, computational intelligence for the IoT and Smart-X technologies.