{"title":"ResNet-18对各种激活函数进行图像分类的比较分析","authors":"Gaurav Pandey, S. Srivastava","doi":"10.1109/ICICT57646.2023.10134464","DOIUrl":null,"url":null,"abstract":"Deep neural network and Machine learning are a latest emerging concept in the field of data science. Due to multi-layer hierarchical feature extraction in conjunction with control variables like number of hidden layers, activation functions, and variable parameters like learning rates, initial weights, and decay functions, deep network models perform better than machine learning techniques. While most of these parameter control the learning dynamics or complexity of representation a neural network can deal with, it is only activation function which introduces non-linearity in a network and current state of activation function poses multiple challenges to both practitioners and researchers some of which are: •Vanishing & Exploding gradients during back-propagation •Zero-mean and range of outputs •Compute complexity of function •Predictive performance Due to this reason our objective in current work in focused to explain with reasoning and experiments the landscape of activation functions available. According to a recent study, we have enough cutting-edge activation functions to modify the architecture of the well-known deep network model. Building on top of widely adopted ResNet-18 network architecture in this study. Subsequently, we evaluate the effectiveness of ResNet-18 for image classification using various activation functions.","PeriodicalId":126489,"journal":{"name":"2023 International Conference on Inventive Computation Technologies (ICICT)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"ResNet-18 comparative analysis of various activation functions for image classification\",\"authors\":\"Gaurav Pandey, S. Srivastava\",\"doi\":\"10.1109/ICICT57646.2023.10134464\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Deep neural network and Machine learning are a latest emerging concept in the field of data science. Due to multi-layer hierarchical feature extraction in conjunction with control variables like number of hidden layers, activation functions, and variable parameters like learning rates, initial weights, and decay functions, deep network models perform better than machine learning techniques. While most of these parameter control the learning dynamics or complexity of representation a neural network can deal with, it is only activation function which introduces non-linearity in a network and current state of activation function poses multiple challenges to both practitioners and researchers some of which are: •Vanishing & Exploding gradients during back-propagation •Zero-mean and range of outputs •Compute complexity of function •Predictive performance Due to this reason our objective in current work in focused to explain with reasoning and experiments the landscape of activation functions available. According to a recent study, we have enough cutting-edge activation functions to modify the architecture of the well-known deep network model. Building on top of widely adopted ResNet-18 network architecture in this study. Subsequently, we evaluate the effectiveness of ResNet-18 for image classification using various activation functions.\",\"PeriodicalId\":126489,\"journal\":{\"name\":\"2023 International Conference on Inventive Computation Technologies (ICICT)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-04-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 International Conference on Inventive Computation Technologies (ICICT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICICT57646.2023.10134464\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference on Inventive Computation Technologies (ICICT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICICT57646.2023.10134464","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
ResNet-18 comparative analysis of various activation functions for image classification
Deep neural network and Machine learning are a latest emerging concept in the field of data science. Due to multi-layer hierarchical feature extraction in conjunction with control variables like number of hidden layers, activation functions, and variable parameters like learning rates, initial weights, and decay functions, deep network models perform better than machine learning techniques. While most of these parameter control the learning dynamics or complexity of representation a neural network can deal with, it is only activation function which introduces non-linearity in a network and current state of activation function poses multiple challenges to both practitioners and researchers some of which are: •Vanishing & Exploding gradients during back-propagation •Zero-mean and range of outputs •Compute complexity of function •Predictive performance Due to this reason our objective in current work in focused to explain with reasoning and experiments the landscape of activation functions available. According to a recent study, we have enough cutting-edge activation functions to modify the architecture of the well-known deep network model. Building on top of widely adopted ResNet-18 network architecture in this study. Subsequently, we evaluate the effectiveness of ResNet-18 for image classification using various activation functions.