利用两个新激活函数:modExp 和 modExpm 提高神经网络性能

IF 1 Q4 OPTICS
Heena Kalim, Anuradha Chug, Amit Prakash Singh
{"title":"利用两个新激活函数:modExp 和 modExpm 提高神经网络性能","authors":"Heena Kalim,&nbsp;Anuradha Chug,&nbsp;Amit Prakash Singh","doi":"10.3103/S1060992X24700152","DOIUrl":null,"url":null,"abstract":"<p>The paper introduces two novel activation functions known as modExp and modExp<sub>m</sub>. The activation functions possess several desirable properties, such as being continuously differentiable, bounded, smooth, and non-monotonic. Our studies have shown that modExp and modExp<sub>m</sub> consistently outperform ReLU and other activation functions across a range of challenging datasets and complex models. Initially, the experiments involve training and classifying using a multi-layer perceptron (MLP) on benchmark data sets like the Diagnostic Wisconsin Breast Cancer and Iris Flower datasets. Both modExp and modExp<sub>m</sub> demonstrate impressive performance, with modExp achieving 94.15 and 95.56% and modExp<sub>m</sub> achieving 94.15 and 95.56% respectively, when compared to ReLU, ELU, Tanh, Mish, Softsign, Leaky ReLU, and TanhExp. In addition, a series of experiments were carried out on five different depths of deeper neural networks, ranging from five to eight layers, using MNIST datasets. The modExp<sub>m</sub> activation function demonstrated superior performance accuracy on various neural network configurations, achieving 95.56, 95.43, 94.72, 95.14, and 95.61% on wider 5 layers, slimmer 5 layers, 6 layers, 7 layers, and 8 layers respectively. The modExp activation function also performed well, achieving the second highest accuracy of 95.42, 94.33, 94.76, 95.06, and 95.37% on the same network configurations, outperforming ReLU, ELU, Tanh, Mish, Softsign, Leaky ReLU, and TanhExp. The results of the statistical feature measures show that both activation functions have the highest mean accuracy, the lowest standard deviation, the lowest Root Mean squared Error, the lowest variance, and the lowest Mean squared Error. According to the experiment, both functions converge more quickly than ReLU, which is a significant advantage in Neural network learning.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"33 3","pages":"286 - 301"},"PeriodicalIF":1.0000,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Enhancement of Neural Network Performance with the Use of Two Novel Activation Functions: modExp and modExpm\",\"authors\":\"Heena Kalim,&nbsp;Anuradha Chug,&nbsp;Amit Prakash Singh\",\"doi\":\"10.3103/S1060992X24700152\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>The paper introduces two novel activation functions known as modExp and modExp<sub>m</sub>. The activation functions possess several desirable properties, such as being continuously differentiable, bounded, smooth, and non-monotonic. Our studies have shown that modExp and modExp<sub>m</sub> consistently outperform ReLU and other activation functions across a range of challenging datasets and complex models. Initially, the experiments involve training and classifying using a multi-layer perceptron (MLP) on benchmark data sets like the Diagnostic Wisconsin Breast Cancer and Iris Flower datasets. Both modExp and modExp<sub>m</sub> demonstrate impressive performance, with modExp achieving 94.15 and 95.56% and modExp<sub>m</sub> achieving 94.15 and 95.56% respectively, when compared to ReLU, ELU, Tanh, Mish, Softsign, Leaky ReLU, and TanhExp. In addition, a series of experiments were carried out on five different depths of deeper neural networks, ranging from five to eight layers, using MNIST datasets. The modExp<sub>m</sub> activation function demonstrated superior performance accuracy on various neural network configurations, achieving 95.56, 95.43, 94.72, 95.14, and 95.61% on wider 5 layers, slimmer 5 layers, 6 layers, 7 layers, and 8 layers respectively. The modExp activation function also performed well, achieving the second highest accuracy of 95.42, 94.33, 94.76, 95.06, and 95.37% on the same network configurations, outperforming ReLU, ELU, Tanh, Mish, Softsign, Leaky ReLU, and TanhExp. The results of the statistical feature measures show that both activation functions have the highest mean accuracy, the lowest standard deviation, the lowest Root Mean squared Error, the lowest variance, and the lowest Mean squared Error. According to the experiment, both functions converge more quickly than ReLU, which is a significant advantage in Neural network learning.</p>\",\"PeriodicalId\":721,\"journal\":{\"name\":\"Optical Memory and Neural Networks\",\"volume\":\"33 3\",\"pages\":\"286 - 301\"},\"PeriodicalIF\":1.0000,\"publicationDate\":\"2024-09-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Optical Memory and Neural Networks\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://link.springer.com/article/10.3103/S1060992X24700152\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"OPTICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Optical Memory and Neural Networks","FirstCategoryId":"1085","ListUrlMain":"https://link.springer.com/article/10.3103/S1060992X24700152","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"OPTICS","Score":null,"Total":0}
引用次数: 0

摘要

本文介绍了两个新颖的激活函数,即 modExp 和 modExpm。这两个激活函数具有几个理想的特性,如连续可微、有界、平滑和非单调性。我们的研究表明,在一系列具有挑战性的数据集和复杂模型中,modExp 和 modExpm 的表现始终优于 ReLU 和其他激活函数。最初,实验涉及在基准数据集(如威斯康星州乳腺癌诊断数据集和鸢尾花数据集)上使用多层感知器(MLP)进行训练和分类。与ReLU、ELU、Tanh、Mish、Softsign、Leaky ReLU和TanhExp相比,modExp和modExpm都表现出令人印象深刻的性能,modExp分别达到94.15%和95.56%,modExpm分别达到94.15%和95.56%。 此外,还使用MNIST数据集对五到八层不同深度的深度神经网络进行了一系列实验。modExpm 激活函数在各种神经网络配置上都表现出了卓越的准确性,在较宽的 5 层、较窄的 5 层、6 层、7 层和 8 层上分别达到了 95.56%、95.43%、94.72%、95.14% 和 95.61%。modExp 激活函数也表现出色,在相同的网络配置下分别达到了 95.42%、94.33%、94.76%、95.06% 和 95.37% 的第二高准确率,优于 ReLU、ELU、Tanh、Mish、Softsign、Leaky ReLU 和 TanhExp。统计特征测量结果表明,这两种激活函数的平均精度最高、标准差最小、均方根误差最小、方差最小、均方误差最小。根据实验结果,这两个函数的收敛速度都比 ReLU 快,这在神经网络学习中是一个显著的优势。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

Enhancement of Neural Network Performance with the Use of Two Novel Activation Functions: modExp and modExpm

Enhancement of Neural Network Performance with the Use of Two Novel Activation Functions: modExp and modExpm

The paper introduces two novel activation functions known as modExp and modExpm. The activation functions possess several desirable properties, such as being continuously differentiable, bounded, smooth, and non-monotonic. Our studies have shown that modExp and modExpm consistently outperform ReLU and other activation functions across a range of challenging datasets and complex models. Initially, the experiments involve training and classifying using a multi-layer perceptron (MLP) on benchmark data sets like the Diagnostic Wisconsin Breast Cancer and Iris Flower datasets. Both modExp and modExpm demonstrate impressive performance, with modExp achieving 94.15 and 95.56% and modExpm achieving 94.15 and 95.56% respectively, when compared to ReLU, ELU, Tanh, Mish, Softsign, Leaky ReLU, and TanhExp. In addition, a series of experiments were carried out on five different depths of deeper neural networks, ranging from five to eight layers, using MNIST datasets. The modExpm activation function demonstrated superior performance accuracy on various neural network configurations, achieving 95.56, 95.43, 94.72, 95.14, and 95.61% on wider 5 layers, slimmer 5 layers, 6 layers, 7 layers, and 8 layers respectively. The modExp activation function also performed well, achieving the second highest accuracy of 95.42, 94.33, 94.76, 95.06, and 95.37% on the same network configurations, outperforming ReLU, ELU, Tanh, Mish, Softsign, Leaky ReLU, and TanhExp. The results of the statistical feature measures show that both activation functions have the highest mean accuracy, the lowest standard deviation, the lowest Root Mean squared Error, the lowest variance, and the lowest Mean squared Error. According to the experiment, both functions converge more quickly than ReLU, which is a significant advantage in Neural network learning.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
1.50
自引率
11.10%
发文量
25
期刊介绍: The journal covers a wide range of issues in information optics such as optical memory, mechanisms for optical data recording and processing, photosensitive materials, optical, optoelectronic and holographic nanostructures, and many other related topics. Papers on memory systems using holographic and biological structures and concepts of brain operation are also included. The journal pays particular attention to research in the field of neural net systems that may lead to a new generation of computional technologies by endowing them with intelligence.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信