Enhancing CNN model classification performance through RGB angle rotation method

Yahya Dogan, Cuneyt Ozdemir, Yılmaz Kaya
{"title":"Enhancing CNN model classification performance through RGB angle rotation method","authors":"Yahya Dogan, Cuneyt Ozdemir, Yılmaz Kaya","doi":"10.1007/s00521-024-10232-z","DOIUrl":null,"url":null,"abstract":"<p>In recent years, convolutional neural networks have significantly advanced the field of computer vision by automatically extracting features from image data. CNNs enable the modeling of complex and abstract image features using learnable filters, eliminating the need for manual feature extraction. However, combining feature maps obtained from CNNs with different approaches can lead to more complex and interpretable inferences, thereby enhancing model performance and generalizability. In this study, we propose a new method called RGB angle rotation to effectively obtain feature maps from RGB images. Our method rotates color channels at different angles and uses the angle information between channels to generate new feature maps. We then investigate the effects of integrating models trained with these feature maps into an ensemble architecture. Experimental results on the CIFAR-10 dataset show that using the proposed method in the ensemble model results in performance increases of 9.10 and 8.42% for the B and R channels, respectively, compared to the original model, while the effect of the G channel is very limited. For the CIFAR-100 dataset, the proposed method resulted in a 17.09% improvement in ensemble model performance for the R channel, a 5.06% increase for the B channel, and no significant improvement for the G channel compared to the original model. Additionally, we compared our method with traditional feature extraction methods like scale-invariant feature transform and local binary pattern and observed higher performance. In conclusion, it has been observed that the proposed RGB angle rotation method significantly impacts model performance.</p>","PeriodicalId":18925,"journal":{"name":"Neural Computing and Applications","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Computing and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s00521-024-10232-z","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In recent years, convolutional neural networks have significantly advanced the field of computer vision by automatically extracting features from image data. CNNs enable the modeling of complex and abstract image features using learnable filters, eliminating the need for manual feature extraction. However, combining feature maps obtained from CNNs with different approaches can lead to more complex and interpretable inferences, thereby enhancing model performance and generalizability. In this study, we propose a new method called RGB angle rotation to effectively obtain feature maps from RGB images. Our method rotates color channels at different angles and uses the angle information between channels to generate new feature maps. We then investigate the effects of integrating models trained with these feature maps into an ensemble architecture. Experimental results on the CIFAR-10 dataset show that using the proposed method in the ensemble model results in performance increases of 9.10 and 8.42% for the B and R channels, respectively, compared to the original model, while the effect of the G channel is very limited. For the CIFAR-100 dataset, the proposed method resulted in a 17.09% improvement in ensemble model performance for the R channel, a 5.06% increase for the B channel, and no significant improvement for the G channel compared to the original model. Additionally, we compared our method with traditional feature extraction methods like scale-invariant feature transform and local binary pattern and observed higher performance. In conclusion, it has been observed that the proposed RGB angle rotation method significantly impacts model performance.

Abstract Image

通过 RGB 角度旋转方法提高 CNN 模型分类性能
近年来,卷积神经网络通过从图像数据中自动提取特征,极大地推动了计算机视觉领域的发展。卷积神经网络能够利用可学习的滤波器对复杂抽象的图像特征进行建模,从而消除了人工特征提取的需要。然而,将从 CNN 中获得的特征图与不同的方法相结合,可以得到更复杂、更可解释的推论,从而提高模型的性能和普适性。在本研究中,我们提出了一种名为 RGB 角度旋转的新方法,以有效地从 RGB 图像中获取特征图。我们的方法以不同的角度旋转颜色通道,并利用通道间的角度信息生成新的特征图。然后,我们研究了将使用这些特征图训练的模型集成到集合架构中的效果。在 CIFAR-10 数据集上的实验结果表明,与原始模型相比,在集合模型中使用所提出的方法,B 和 R 信道的性能分别提高了 9.10% 和 8.42%,而 G 信道的影响则非常有限。对于 CIFAR-100 数据集,与原始模型相比,建议方法使 R 信道的集合模型性能提高了 17.09%,B 信道提高了 5.06%,而 G 信道没有显著提高。此外,我们还将我们的方法与传统的特征提取方法(如尺度不变特征变换和局部二进制模式)进行了比较,发现我们的方法性能更高。总之,我们发现所提出的 RGB 角度旋转方法对模型性能有显著影响。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信