{"title":"Continuous Action Learning Automata Optimizer for training Artificial Neural Networks","authors":"J. Lindsay, S. Givigi","doi":"10.1109/SysCon53073.2023.10131086","DOIUrl":null,"url":null,"abstract":"This paper introduces a Continuous-Action Learning Automata (CALA) game optimizer that provides a generalized way to use a game of CALA agents to train Artificial Neural Networks (ANNs) and Deep ANNs. This method uses both game theory and learning automata, which makes it a computationally efficient method when compared against other non-gradient and non-back propagation methods. Since the CALA game optimizer does not use gradients or back propagation, issues such as the vanishing gradient problem do not manifest, which allows for the use of multiple activation functions such as sigmoid or tanh even in a Deep ANN.","PeriodicalId":169296,"journal":{"name":"2023 IEEE International Systems Conference (SysCon)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE International Systems Conference (SysCon)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SysCon53073.2023.10131086","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper introduces a Continuous-Action Learning Automata (CALA) game optimizer that provides a generalized way to use a game of CALA agents to train Artificial Neural Networks (ANNs) and Deep ANNs. This method uses both game theory and learning automata, which makes it a computationally efficient method when compared against other non-gradient and non-back propagation methods. Since the CALA game optimizer does not use gradients or back propagation, issues such as the vanishing gradient problem do not manifest, which allows for the use of multiple activation functions such as sigmoid or tanh even in a Deep ANN.