通过正则化参数的计算优化改进深度随机向量泛函链路网络

IF 8 2区 计算机科学 Q1 AUTOMATION & CONTROL SYSTEMS
Chinnamuthu Subramani , Ravi Prasad K. Jagannath , Venkatanareshbabu Kuppili
{"title":"通过正则化参数的计算优化改进深度随机向量泛函链路网络","authors":"Chinnamuthu Subramani ,&nbsp;Ravi Prasad K. Jagannath ,&nbsp;Venkatanareshbabu Kuppili","doi":"10.1016/j.engappai.2025.110389","DOIUrl":null,"url":null,"abstract":"<div><div>Deep Random Vector Functional Link (dRVFL) networks are a class of randomization-based deep neural networks known for their rapid learning capabilities and universal approximation potential. Despite these advantages, the dRVFL model faces substantial computational and memory challenges as training data, hidden layers, or input feature dimensions grow. These issues are intensified by regularization, as optimal performance requires repeatedly solving large-scale linear systems to tune the regularization parameter. Traditional matrix inversion methods for this task are computationally intensive, memory-demanding, and prone to numerical instability. This study introduces a computationally efficient two-stage approach to address these limitations. In the first stage, a randomized algorithm-based low-rank approximation computes a compressed representation of the feature matrix in <span><math><mrow><mi>O</mi><mrow><mo>(</mo><mi>m</mi><mi>n</mi><mi>k</mi><mo>)</mo></mrow></mrow></math></span> for an <span><math><mrow><mi>m</mi><mo>×</mo><mi>n</mi></mrow></math></span> matrix (<span><math><mrow><mi>k</mi><mo>≪</mo><mo>min</mo><mrow><mo>(</mo><mi>m</mi><mo>,</mo><mi>n</mi><mo>)</mo></mrow></mrow></math></span>), requiring only two passes over the data, significantly reducing costs. In the second stage, the derived solution is used to optimize <span><math><mi>λ</mi></math></span> via the Golden Section search and Menger curvature-based technique, enabling efficient exploration of the continuous domain with minimal function evaluations. Furthermore, the theoretical framework derives an upper bound on the error of the approximated regularized solution, combining deterministic and probabilistic insights. The proposed method, validated on multiple classification datasets, demonstrates up to 6% improvement in accuracy compared to state-of-the-art feedforward neural network algorithms. Statistical evaluations further confirm its robustness and superior performance over seven baseline algorithms. Overall, the proposed method enhances the scalability, robustness, and efficiency of dRVFL for large-scale applications.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"148 ","pages":"Article 110389"},"PeriodicalIF":8.0000,"publicationDate":"2025-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Improving Deep Random Vector Functional Link Networks through computational optimization of regularization parameters\",\"authors\":\"Chinnamuthu Subramani ,&nbsp;Ravi Prasad K. Jagannath ,&nbsp;Venkatanareshbabu Kuppili\",\"doi\":\"10.1016/j.engappai.2025.110389\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Deep Random Vector Functional Link (dRVFL) networks are a class of randomization-based deep neural networks known for their rapid learning capabilities and universal approximation potential. Despite these advantages, the dRVFL model faces substantial computational and memory challenges as training data, hidden layers, or input feature dimensions grow. These issues are intensified by regularization, as optimal performance requires repeatedly solving large-scale linear systems to tune the regularization parameter. Traditional matrix inversion methods for this task are computationally intensive, memory-demanding, and prone to numerical instability. This study introduces a computationally efficient two-stage approach to address these limitations. In the first stage, a randomized algorithm-based low-rank approximation computes a compressed representation of the feature matrix in <span><math><mrow><mi>O</mi><mrow><mo>(</mo><mi>m</mi><mi>n</mi><mi>k</mi><mo>)</mo></mrow></mrow></math></span> for an <span><math><mrow><mi>m</mi><mo>×</mo><mi>n</mi></mrow></math></span> matrix (<span><math><mrow><mi>k</mi><mo>≪</mo><mo>min</mo><mrow><mo>(</mo><mi>m</mi><mo>,</mo><mi>n</mi><mo>)</mo></mrow></mrow></math></span>), requiring only two passes over the data, significantly reducing costs. In the second stage, the derived solution is used to optimize <span><math><mi>λ</mi></math></span> via the Golden Section search and Menger curvature-based technique, enabling efficient exploration of the continuous domain with minimal function evaluations. Furthermore, the theoretical framework derives an upper bound on the error of the approximated regularized solution, combining deterministic and probabilistic insights. The proposed method, validated on multiple classification datasets, demonstrates up to 6% improvement in accuracy compared to state-of-the-art feedforward neural network algorithms. Statistical evaluations further confirm its robustness and superior performance over seven baseline algorithms. Overall, the proposed method enhances the scalability, robustness, and efficiency of dRVFL for large-scale applications.</div></div>\",\"PeriodicalId\":50523,\"journal\":{\"name\":\"Engineering Applications of Artificial Intelligence\",\"volume\":\"148 \",\"pages\":\"Article 110389\"},\"PeriodicalIF\":8.0000,\"publicationDate\":\"2025-03-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Engineering Applications of Artificial Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0952197625003896\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625003896","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

摘要

深度随机向量功能链接(dRVFL)网络是一类基于随机化的深度神经网络,以其快速学习能力和通用逼近潜力而闻名。尽管有这些优势,但随着训练数据、隐藏层或输入特征维度的增长,dRVFL模型面临着大量的计算和内存挑战。这些问题由于正则化而加剧,因为优化性能需要反复求解大规模线性系统来调整正则化参数。传统的矩阵反演方法计算量大,内存要求高,并且容易出现数值不稳定性。本研究引入了一种计算效率高的两阶段方法来解决这些限制。在第一阶段,基于随机算法的低秩近似值为m×n矩阵(k≪min(m,n))计算O(mnk)中特征矩阵的压缩表示,只需要对数据进行两次遍历,大大降低了成本。在第二阶段,利用导出的解通过黄金分割搜索和基于Menger曲率的技术来优化λ,从而以最小的函数评估实现对连续域的有效探索。此外,理论框架推导了近似正则化解的误差上界,结合了确定性和概率性的见解。该方法在多个分类数据集上进行了验证,与最先进的前馈神经网络算法相比,准确率提高了6%。统计评估进一步证实了其鲁棒性和优于七个基线算法的性能。总体而言,该方法提高了dRVFL在大规模应用中的可扩展性、鲁棒性和效率。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Improving Deep Random Vector Functional Link Networks through computational optimization of regularization parameters
Deep Random Vector Functional Link (dRVFL) networks are a class of randomization-based deep neural networks known for their rapid learning capabilities and universal approximation potential. Despite these advantages, the dRVFL model faces substantial computational and memory challenges as training data, hidden layers, or input feature dimensions grow. These issues are intensified by regularization, as optimal performance requires repeatedly solving large-scale linear systems to tune the regularization parameter. Traditional matrix inversion methods for this task are computationally intensive, memory-demanding, and prone to numerical instability. This study introduces a computationally efficient two-stage approach to address these limitations. In the first stage, a randomized algorithm-based low-rank approximation computes a compressed representation of the feature matrix in O(mnk) for an m×n matrix (kmin(m,n)), requiring only two passes over the data, significantly reducing costs. In the second stage, the derived solution is used to optimize λ via the Golden Section search and Menger curvature-based technique, enabling efficient exploration of the continuous domain with minimal function evaluations. Furthermore, the theoretical framework derives an upper bound on the error of the approximated regularized solution, combining deterministic and probabilistic insights. The proposed method, validated on multiple classification datasets, demonstrates up to 6% improvement in accuracy compared to state-of-the-art feedforward neural network algorithms. Statistical evaluations further confirm its robustness and superior performance over seven baseline algorithms. Overall, the proposed method enhances the scalability, robustness, and efficiency of dRVFL for large-scale applications.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Engineering Applications of Artificial Intelligence
Engineering Applications of Artificial Intelligence 工程技术-工程:电子与电气
CiteScore
9.60
自引率
10.00%
发文量
505
审稿时长
68 days
期刊介绍: Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信