Mighty Abra Ayidzoe, Yongbin Yu, Patrick Kwabena Mensah, Jingye Cai, F. U. Bawah, O. Nyarko-Boateng
{"title":"Patch-Based Capsule Network for Complex Images","authors":"Mighty Abra Ayidzoe, Yongbin Yu, Patrick Kwabena Mensah, Jingye Cai, F. U. Bawah, O. Nyarko-Boateng","doi":"10.1109/ICAST52759.2021.9681959","DOIUrl":null,"url":null,"abstract":"Capsule Networks are neural networks that have the advantage of learning spatial and hierarchical information from data. They can learn and extract knowledge from smaller datasets (unlike other neural network algorithms); however, they perform poorly on complex and low-resolution images due to a problem called “crowding.” Crowding is attributed to the CapsNets’ attempt to read every object in an image (including background objects), resulting in poor performance. Therefore, this paper proposes a patch-based capsule network and a new squash function (power-B) to decompose an input image into smaller parts enabling the model to focus more on the relevant regions of interest. Experimental results show that the proposed model has efficient feature extraction capabilities, reduced computational time, and a fewer trainable number of parameters. The model’s performance is comparable to the state-of-the-art capsule network models by achieving overall recognition accuracies of 94.62%, 75.68%, and 92.82% for fashion-MNIST, CIFAR 10, and polyp datasets, respectively.","PeriodicalId":434382,"journal":{"name":"2021 IEEE 8th International Conference on Adaptive Science and Technology (ICAST)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 8th International Conference on Adaptive Science and Technology (ICAST)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAST52759.2021.9681959","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Capsule Networks are neural networks that have the advantage of learning spatial and hierarchical information from data. They can learn and extract knowledge from smaller datasets (unlike other neural network algorithms); however, they perform poorly on complex and low-resolution images due to a problem called “crowding.” Crowding is attributed to the CapsNets’ attempt to read every object in an image (including background objects), resulting in poor performance. Therefore, this paper proposes a patch-based capsule network and a new squash function (power-B) to decompose an input image into smaller parts enabling the model to focus more on the relevant regions of interest. Experimental results show that the proposed model has efficient feature extraction capabilities, reduced computational time, and a fewer trainable number of parameters. The model’s performance is comparable to the state-of-the-art capsule network models by achieving overall recognition accuracies of 94.62%, 75.68%, and 92.82% for fashion-MNIST, CIFAR 10, and polyp datasets, respectively.