Merve Güney Duman , Sibel Koparal , Neşe Ömür , Alp Ertürk , Erchan Aptoula
{"title":"自适应双参数激活函数","authors":"Merve Güney Duman , Sibel Koparal , Neşe Ömür , Alp Ertürk , Erchan Aptoula","doi":"10.1016/j.dsp.2025.105579","DOIUrl":null,"url":null,"abstract":"<div><div>Activation functions are critical components of neural networks, introducing the necessary nonlinearity for learning complex data relationships. While widely used functions such as ReLU and its variants have demonstrated notable success, they still suffer from limitations such as vanishing gradients, dead neurons, and limited adaptability at various degrees. This paper proposes two novel differentiable double-parameter activation functions (AdLU<span><math><msub><mrow></mrow><mn>1</mn></msub></math></span> and AdLU<span><math><msub><mrow></mrow><mn>2</mn></msub></math></span>) designed to address these challenges. They incorporate tunable parameters to optimize gradient flow and enhance adaptability. Evaluations on benchmark datasets, MNIST, FMNIST, USPS, and CIFAR-10, using ResNet-18 and ResNet-50 architectures, demonstrate that the proposed functions consistently achieve high classification accuracy. Notably, AdLU<span><math><msub><mrow></mrow><mn>1</mn></msub></math></span> improves accuracy by up to 5.5 % compared to ReLU, particularly in deeper architectures and more complex datasets. While introducing some computational overhead, their performance gains establish them as competitive alternatives to both traditional and modern activation functions.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"168 ","pages":"Article 105579"},"PeriodicalIF":3.0000,"publicationDate":"2025-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"AdLU: Adaptive double parametric activation functions\",\"authors\":\"Merve Güney Duman , Sibel Koparal , Neşe Ömür , Alp Ertürk , Erchan Aptoula\",\"doi\":\"10.1016/j.dsp.2025.105579\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Activation functions are critical components of neural networks, introducing the necessary nonlinearity for learning complex data relationships. While widely used functions such as ReLU and its variants have demonstrated notable success, they still suffer from limitations such as vanishing gradients, dead neurons, and limited adaptability at various degrees. This paper proposes two novel differentiable double-parameter activation functions (AdLU<span><math><msub><mrow></mrow><mn>1</mn></msub></math></span> and AdLU<span><math><msub><mrow></mrow><mn>2</mn></msub></math></span>) designed to address these challenges. They incorporate tunable parameters to optimize gradient flow and enhance adaptability. Evaluations on benchmark datasets, MNIST, FMNIST, USPS, and CIFAR-10, using ResNet-18 and ResNet-50 architectures, demonstrate that the proposed functions consistently achieve high classification accuracy. Notably, AdLU<span><math><msub><mrow></mrow><mn>1</mn></msub></math></span> improves accuracy by up to 5.5 % compared to ReLU, particularly in deeper architectures and more complex datasets. While introducing some computational overhead, their performance gains establish them as competitive alternatives to both traditional and modern activation functions.</div></div>\",\"PeriodicalId\":51011,\"journal\":{\"name\":\"Digital Signal Processing\",\"volume\":\"168 \",\"pages\":\"Article 105579\"},\"PeriodicalIF\":3.0000,\"publicationDate\":\"2025-09-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Digital Signal Processing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1051200425006013\",\"RegionNum\":3,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Digital Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1051200425006013","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Activation functions are critical components of neural networks, introducing the necessary nonlinearity for learning complex data relationships. While widely used functions such as ReLU and its variants have demonstrated notable success, they still suffer from limitations such as vanishing gradients, dead neurons, and limited adaptability at various degrees. This paper proposes two novel differentiable double-parameter activation functions (AdLU and AdLU) designed to address these challenges. They incorporate tunable parameters to optimize gradient flow and enhance adaptability. Evaluations on benchmark datasets, MNIST, FMNIST, USPS, and CIFAR-10, using ResNet-18 and ResNet-50 architectures, demonstrate that the proposed functions consistently achieve high classification accuracy. Notably, AdLU improves accuracy by up to 5.5 % compared to ReLU, particularly in deeper architectures and more complex datasets. While introducing some computational overhead, their performance gains establish them as competitive alternatives to both traditional and modern activation functions.
期刊介绍:
Digital Signal Processing: A Review Journal is one of the oldest and most established journals in the field of signal processing yet it aims to be the most innovative. The Journal invites top quality research articles at the frontiers of research in all aspects of signal processing. Our objective is to provide a platform for the publication of ground-breaking research in signal processing with both academic and industrial appeal.
The journal has a special emphasis on statistical signal processing methodology such as Bayesian signal processing, and encourages articles on emerging applications of signal processing such as:
• big data• machine learning• internet of things• information security• systems biology and computational biology,• financial time series analysis,• autonomous vehicles,• quantum computing,• neuromorphic engineering,• human-computer interaction and intelligent user interfaces,• environmental signal processing,• geophysical signal processing including seismic signal processing,• chemioinformatics and bioinformatics,• audio, visual and performance arts,• disaster management and prevention,• renewable energy,