AdLU: Adaptive double parametric activation functions

IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC
Merve Güney Duman , Sibel Koparal , Neşe Ömür , Alp Ertürk , Erchan Aptoula
{"title":"AdLU: Adaptive double parametric activation functions","authors":"Merve Güney Duman ,&nbsp;Sibel Koparal ,&nbsp;Neşe Ömür ,&nbsp;Alp Ertürk ,&nbsp;Erchan Aptoula","doi":"10.1016/j.dsp.2025.105579","DOIUrl":null,"url":null,"abstract":"<div><div>Activation functions are critical components of neural networks, introducing the necessary nonlinearity for learning complex data relationships. While widely used functions such as ReLU and its variants have demonstrated notable success, they still suffer from limitations such as vanishing gradients, dead neurons, and limited adaptability at various degrees. This paper proposes two novel differentiable double-parameter activation functions (AdLU<span><math><msub><mrow></mrow><mn>1</mn></msub></math></span> and AdLU<span><math><msub><mrow></mrow><mn>2</mn></msub></math></span>) designed to address these challenges. They incorporate tunable parameters to optimize gradient flow and enhance adaptability. Evaluations on benchmark datasets, MNIST, FMNIST, USPS, and CIFAR-10, using ResNet-18 and ResNet-50 architectures, demonstrate that the proposed functions consistently achieve high classification accuracy. Notably, AdLU<span><math><msub><mrow></mrow><mn>1</mn></msub></math></span> improves accuracy by up to 5.5 % compared to ReLU, particularly in deeper architectures and more complex datasets. While introducing some computational overhead, their performance gains establish them as competitive alternatives to both traditional and modern activation functions.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"168 ","pages":"Article 105579"},"PeriodicalIF":3.0000,"publicationDate":"2025-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Digital Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1051200425006013","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

Activation functions are critical components of neural networks, introducing the necessary nonlinearity for learning complex data relationships. While widely used functions such as ReLU and its variants have demonstrated notable success, they still suffer from limitations such as vanishing gradients, dead neurons, and limited adaptability at various degrees. This paper proposes two novel differentiable double-parameter activation functions (AdLU1 and AdLU2) designed to address these challenges. They incorporate tunable parameters to optimize gradient flow and enhance adaptability. Evaluations on benchmark datasets, MNIST, FMNIST, USPS, and CIFAR-10, using ResNet-18 and ResNet-50 architectures, demonstrate that the proposed functions consistently achieve high classification accuracy. Notably, AdLU1 improves accuracy by up to 5.5 % compared to ReLU, particularly in deeper architectures and more complex datasets. While introducing some computational overhead, their performance gains establish them as competitive alternatives to both traditional and modern activation functions.
自适应双参数激活函数
激活函数是神经网络的重要组成部分,它引入了学习复杂数据关系所必需的非线性。虽然广泛使用的函数(如ReLU及其变体)已经取得了显著的成功,但它们仍然存在诸如梯度消失、死神经元和不同程度的有限适应性等局限性。本文提出了两个新的可微双参数激活函数(AdLU1和AdLU2)来解决这些问题。它们采用可调参数来优化梯度流,增强适应性。在使用ResNet-18和ResNet-50架构的基准数据集MNIST、FMNIST、USPS和CIFAR-10上的评估表明,所提出的函数始终保持较高的分类精度。值得注意的是,与ReLU相比,AdLU1的准确率提高了5.5%,特别是在更深层次的架构和更复杂的数据集中。虽然引入了一些计算开销,但它们的性能增益使它们成为传统和现代激活函数的竞争性替代品。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Digital Signal Processing
Digital Signal Processing 工程技术-工程:电子与电气
CiteScore
5.30
自引率
17.20%
发文量
435
审稿时长
66 days
期刊介绍: Digital Signal Processing: A Review Journal is one of the oldest and most established journals in the field of signal processing yet it aims to be the most innovative. The Journal invites top quality research articles at the frontiers of research in all aspects of signal processing. Our objective is to provide a platform for the publication of ground-breaking research in signal processing with both academic and industrial appeal. The journal has a special emphasis on statistical signal processing methodology such as Bayesian signal processing, and encourages articles on emerging applications of signal processing such as: • big data• machine learning• internet of things• information security• systems biology and computational biology,• financial time series analysis,• autonomous vehicles,• quantum computing,• neuromorphic engineering,• human-computer interaction and intelligent user interfaces,• environmental signal processing,• geophysical signal processing including seismic signal processing,• chemioinformatics and bioinformatics,• audio, visual and performance arts,• disaster management and prevention,• renewable energy,
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信