{"title":"The New Activation Function for Complex Valued Neural Networks: Complex Swish Function","authors":"M. Celebi, M. Ceylan","doi":"10.36287/setsci.4.6.050","DOIUrl":null,"url":null,"abstract":"Complex-valued artificial neural network (CVANN) has been developed to process data with complex numbers directly. Weights, threshold, inputs and outputs are all complex numbers in the CVANN. The convergence of the CVANN back propagation algorithm depends on some factors such as selection of appropriate activation function, threshold values, initial weights and normalization of data. The most important of these factors is the selection of the appropriate activation function. The selection of activation function determines the convergence and general formation characteristics of the complex back propagation algorithm. In this study, the swish activation function discovered by Google researchers Prajit Ramachandra, Barret Zoph and Quoc V. Le is discussed in the complex domain. Swish activation function, which gives good results in real plane, has been studied in the complex plane. We have compared the performance of swish activation functions on the complex XOR and symmetry problems with other known activation functions. The simulations’ results show that the proposed network using swish activation function, gives the best results when compared to other networks using the traditional complex logarithmic sigmoid and tangent sigmoid activation functions.","PeriodicalId":6817,"journal":{"name":"4th International Symposium on Innovative Approaches in Engineering and Natural Sciences Proceedings","volume":"49 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2019-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"4th International Symposium on Innovative Approaches in Engineering and Natural Sciences Proceedings","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.36287/setsci.4.6.050","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
Complex-valued artificial neural network (CVANN) has been developed to process data with complex numbers directly. Weights, threshold, inputs and outputs are all complex numbers in the CVANN. The convergence of the CVANN back propagation algorithm depends on some factors such as selection of appropriate activation function, threshold values, initial weights and normalization of data. The most important of these factors is the selection of the appropriate activation function. The selection of activation function determines the convergence and general formation characteristics of the complex back propagation algorithm. In this study, the swish activation function discovered by Google researchers Prajit Ramachandra, Barret Zoph and Quoc V. Le is discussed in the complex domain. Swish activation function, which gives good results in real plane, has been studied in the complex plane. We have compared the performance of swish activation functions on the complex XOR and symmetry problems with other known activation functions. The simulations’ results show that the proposed network using swish activation function, gives the best results when compared to other networks using the traditional complex logarithmic sigmoid and tangent sigmoid activation functions.