Image Retrieval by Fusion of Features from Pre-trained Deep Convolution Neural Networks

Vijayakumar Bhandi, K. S. Sumithra Devi
{"title":"Image Retrieval by Fusion of Features from Pre-trained Deep Convolution Neural Networks","authors":"Vijayakumar Bhandi, K. S. Sumithra Devi","doi":"10.1109/ICATIECE45860.2019.9063814","DOIUrl":null,"url":null,"abstract":"Image retrieval is a challenging problem in computer vision domain. Traditional content based image retrieval (CBIR) systems were built to retrieve images based on low level content representations like color, texture and shape. These domain specific handcrafted features performed well in various image retrieval applications. The choice of image features greatly affects the performance of such systems. Also, one needs deeper understanding of the domain in order to choose right features for image retrieval application. Recent advances in image retrieval focus on creating features which are domain independent. Machine learning can help to learn important representations from images. Convolution neural networks (CNN) are an important class of machine learning models. CNNs can derive high-level multi-scale features from image data. CNNs with deep layers are widely used in image classification problems. Creating a new effective deep CNN model requires huge training time, computing resources and big datasets. There are many deep CNN models like VGG16, ResNet, Alexnet etc., which are pre-trained on huge datasets and model weights are shared for transferring the learnt knowledge to new domains. Pre-trained CNNs can be applied to image retrieval problem by extracting features from fully connected layers of the model before output layer. In this work, two leading pre-trained CNN models VGG16 and ResNet are used to create a CBIR method. Learnt features from these pre-trained models are used to create a fusion feature and use them for image retrieval. The proposed CBIR framework is applied to image retrieval problem in a different domain, satellite images.","PeriodicalId":106496,"journal":{"name":"2019 1st International Conference on Advanced Technologies in Intelligent Control, Environment, Computing & Communication Engineering (ICATIECE)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"12","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 1st International Conference on Advanced Technologies in Intelligent Control, Environment, Computing & Communication Engineering (ICATIECE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICATIECE45860.2019.9063814","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 12

Abstract

Image retrieval is a challenging problem in computer vision domain. Traditional content based image retrieval (CBIR) systems were built to retrieve images based on low level content representations like color, texture and shape. These domain specific handcrafted features performed well in various image retrieval applications. The choice of image features greatly affects the performance of such systems. Also, one needs deeper understanding of the domain in order to choose right features for image retrieval application. Recent advances in image retrieval focus on creating features which are domain independent. Machine learning can help to learn important representations from images. Convolution neural networks (CNN) are an important class of machine learning models. CNNs can derive high-level multi-scale features from image data. CNNs with deep layers are widely used in image classification problems. Creating a new effective deep CNN model requires huge training time, computing resources and big datasets. There are many deep CNN models like VGG16, ResNet, Alexnet etc., which are pre-trained on huge datasets and model weights are shared for transferring the learnt knowledge to new domains. Pre-trained CNNs can be applied to image retrieval problem by extracting features from fully connected layers of the model before output layer. In this work, two leading pre-trained CNN models VGG16 and ResNet are used to create a CBIR method. Learnt features from these pre-trained models are used to create a fusion feature and use them for image retrieval. The proposed CBIR framework is applied to image retrieval problem in a different domain, satellite images.
基于预训练深度卷积神经网络特征融合的图像检索
图像检索是计算机视觉领域的一个具有挑战性的问题。传统的基于内容的图像检索(CBIR)系统是基于颜色、纹理和形状等低级内容表示来检索图像的。这些特定领域的手工特征在各种图像检索应用中表现良好。图像特征的选择极大地影响了系统的性能。此外,为了在图像检索应用中选择正确的特征,需要对图像检索领域有更深入的了解。近年来图像检索的研究进展主要集中在创建与域无关的特征。机器学习可以帮助从图像中学习重要的表征。卷积神经网络(CNN)是一类重要的机器学习模型。cnn可以从图像数据中获得高级别的多尺度特征。深层cnn广泛应用于图像分类问题。创建一个新的有效的深度CNN模型需要大量的训练时间、计算资源和大数据集。有许多深度CNN模型,如VGG16, ResNet, Alexnet等,它们是在庞大的数据集上进行预训练的,并且共享模型权重,以便将学习到的知识转移到新的领域。预训练的cnn可以通过在输出层之前从模型的全连接层中提取特征来应用于图像检索问题。在这项工作中,使用两个领先的预训练CNN模型VGG16和ResNet来创建CBIR方法。从这些预训练模型中学习到的特征用于创建融合特征并将其用于图像检索。提出的CBIR框架应用于不同领域的图像检索问题,即卫星图像。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信