R-CapsNet: An Improvement of Capsule Network for More Complex Data

Lu Luo, Shukai Duan, Lidan Wang
{"title":"R-CapsNet: An Improvement of Capsule Network for More Complex Data","authors":"Lu Luo, Shukai Duan, Lidan Wang","doi":"10.1109/SSCI44817.2019.9003060","DOIUrl":null,"url":null,"abstract":"Convolutional neural networks (CNNs) have achieved the best performance in some fields. However, they still have some defects. CNNs need a lot of images for training; they will lose much information in the pooling layer, which reduces the spatial resolution. Facing such problems, Hinton et al. proposed a capsule network (CapsNet). Although the CapsNet has achieved the best accuracy on MNIST dataset, it has not performed well on Fashion-MNIST, Cifar-10 and other datasets. Naturally, we established an improved version of capsule network (R-CapsNet). Results have shown that when using R-CapsNet model, the loss gets decreased and the accuracy gets improved on FashionMNIST. In the meanwhile, the training parameters are reduced by nearly half. Specifically, it reduces by 4.5M. Comparisons show that our proposed model reports improved accuracy of around 0.56% over the existing state-of-the-art systems in literature. The test accuracy of R-CapsNet model is 1.32% higher than that of the original model. Furthermore, better results have been achieved on Cifar-10 with R-CapsNet model and it has easily increased by 10% compared to CapsNet.","PeriodicalId":6729,"journal":{"name":"2019 IEEE Symposium Series on Computational Intelligence (SSCI)","volume":"56 1","pages":"2124-2129"},"PeriodicalIF":0.0000,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Symposium Series on Computational Intelligence (SSCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SSCI44817.2019.9003060","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Convolutional neural networks (CNNs) have achieved the best performance in some fields. However, they still have some defects. CNNs need a lot of images for training; they will lose much information in the pooling layer, which reduces the spatial resolution. Facing such problems, Hinton et al. proposed a capsule network (CapsNet). Although the CapsNet has achieved the best accuracy on MNIST dataset, it has not performed well on Fashion-MNIST, Cifar-10 and other datasets. Naturally, we established an improved version of capsule network (R-CapsNet). Results have shown that when using R-CapsNet model, the loss gets decreased and the accuracy gets improved on FashionMNIST. In the meanwhile, the training parameters are reduced by nearly half. Specifically, it reduces by 4.5M. Comparisons show that our proposed model reports improved accuracy of around 0.56% over the existing state-of-the-art systems in literature. The test accuracy of R-CapsNet model is 1.32% higher than that of the original model. Furthermore, better results have been achieved on Cifar-10 with R-CapsNet model and it has easily increased by 10% compared to CapsNet.
R-CapsNet:一种针对更复杂数据的胶囊网络改进
卷积神经网络(cnn)在一些领域取得了最好的表现。然而,它们仍然存在一些缺陷。cnn需要大量的图像进行训练;它们会在池化层中丢失大量信息,从而降低空间分辨率。面对这样的问题,Hinton等人提出了胶囊网络(CapsNet)。虽然CapsNet在MNIST数据集上取得了最好的精度,但在Fashion-MNIST、Cifar-10等数据集上表现不佳。自然,我们建立了一个改进版本的胶囊网络(R-CapsNet)。结果表明,使用R-CapsNet模型时,在FashionMNIST上减少了损失,提高了准确率。同时,训练参数减少了近一半。具体来说,它减少了450万。比较表明,我们提出的模型报告的准确性比文献中现有的最先进的系统提高了约0.56%。R-CapsNet模型的测试精度比原模型提高1.32%。此外,使用R-CapsNet模型在Cifar-10上取得了较好的结果,比CapsNet模型轻松提高了10%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信