Unification of popular artificial neural network activation functions

IF 2.5 2区 数学 Q1 MATHEMATICS
Mohammad Mostafanejad
{"title":"Unification of popular artificial neural network activation functions","authors":"Mohammad Mostafanejad","doi":"10.1007/s13540-024-00347-4","DOIUrl":null,"url":null,"abstract":"<p>We present a unified representation of the most popular neural network activation functions. Adopting Mittag-Leffler functions of fractional calculus, we propose a flexible and compact functional form that is able to interpolate between various activation functions and mitigate common problems in training deep neural networks such as vanishing and exploding gradients. The presented gated representation extends the scope of fixed-shape activation functions to their adaptive counterparts whose shape can be learnt from the training data. The derivatives of the proposed functional form can also be expressed in terms of Mittag-Leffler functions making it suitable for backpropagation algorithms. By training an array of neural network architectures of different complexities on various benchmark datasets, we demonstrate that adopting a unified gated representation of activation functions offers a promising and affordable alternative to individual built-in implementations of activation functions in conventional machine learning frameworks.</p>","PeriodicalId":48928,"journal":{"name":"Fractional Calculus and Applied Analysis","volume":"108 1","pages":""},"PeriodicalIF":2.5000,"publicationDate":"2024-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Fractional Calculus and Applied Analysis","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1007/s13540-024-00347-4","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS","Score":null,"Total":0}
引用次数: 0

Abstract

We present a unified representation of the most popular neural network activation functions. Adopting Mittag-Leffler functions of fractional calculus, we propose a flexible and compact functional form that is able to interpolate between various activation functions and mitigate common problems in training deep neural networks such as vanishing and exploding gradients. The presented gated representation extends the scope of fixed-shape activation functions to their adaptive counterparts whose shape can be learnt from the training data. The derivatives of the proposed functional form can also be expressed in terms of Mittag-Leffler functions making it suitable for backpropagation algorithms. By training an array of neural network architectures of different complexities on various benchmark datasets, we demonstrate that adopting a unified gated representation of activation functions offers a promising and affordable alternative to individual built-in implementations of activation functions in conventional machine learning frameworks.

Abstract Image

统一流行的人工神经网络激活函数
我们提出了最流行的神经网络激活函数的统一表示法。通过采用分数微积分的 Mittag-Leffler 函数,我们提出了一种灵活而紧凑的函数形式,它能够在各种激活函数之间进行插值,并缓解深度神经网络训练中的常见问题,如梯度消失和梯度爆炸。所提出的门控表示法将固定形状激活函数的范围扩展到了自适应对应函数,其形状可以从训练数据中学习。所提出的函数形式的导数也可以用 Mittag-Leffler 函数表示,从而使其适用于反向传播算法。通过在各种基准数据集上训练一系列不同复杂度的神经网络架构,我们证明,采用统一的激活函数门控表示法,为传统机器学习框架中激活函数的单个内置实现提供了一种前景广阔且经济实惠的替代方案。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Fractional Calculus and Applied Analysis
Fractional Calculus and Applied Analysis MATHEMATICS, APPLIED-MATHEMATICS, INTERDISCIPLINARY APPLICATIONS
CiteScore
4.70
自引率
16.70%
发文量
101
期刊介绍: Fractional Calculus and Applied Analysis (FCAA, abbreviated in the World databases as Fract. Calc. Appl. Anal. or FRACT CALC APPL ANAL) is a specialized international journal for theory and applications of an important branch of Mathematical Analysis (Calculus) where differentiations and integrations can be of arbitrary non-integer order. The high standards of its contents are guaranteed by the prominent members of Editorial Board and the expertise of invited external reviewers, and proven by the recently achieved high values of impact factor (JIF) and impact rang (SJR), launching the journal to top places of the ranking lists of Thomson Reuters and Scopus.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信