Triplet Deep Hashing with Joint Supervised Loss for Fast Image Retrieval

Mingyong Li, Hongya Wang, Liangliang Wang, Kaixiang Yang, Yingyuan Xiao
{"title":"Triplet Deep Hashing with Joint Supervised Loss for Fast Image Retrieval","authors":"Mingyong Li, Hongya Wang, Liangliang Wang, Kaixiang Yang, Yingyuan Xiao","doi":"10.1109/ICTAI.2019.00090","DOIUrl":null,"url":null,"abstract":"In recent years, the emerging hashing techniques have been successful in large-scale image retrieval. Due to its strong learning ability, deep hashing has become one of the most promising solutions and achieved good results in practice. However, existing deep hashing methods had some limitations, for example, most methods consider only one kind of supervised loss, which leads to insufficient utilization of supervised information. To address this issue, we proposed a Triplet Deep Hashing method with Joint supervised Loss based on convolution neural network (JLTDH) in this work. The proposed JLTDH method combine triplet likelihood loss and linear classification loss, moreover, the triplet supervised label is adopted, which contains richer supervised information than that of pointwise and pairwise label. At the same time, in order to overcome the cubic increase in the number of triplets and make triplet training more effective, we adopt a novel triplet selection method. The whole process is divided into two stages, in the first stage, taking the triplets generated by the triplet selection method as the input of CNN, the three CNNs with shared weights are used for image feature learning, the last layer of the network outputs a preliminary hash code. In the second stage, relying on the hash code of the first stage and the joint loss function, the neural network model is further optimized so that the generated hash code has higher query precision. We perform extensive experiments on three public benchmark datasets CIFAR-10, NUS-WIDE, and MS-COCO. Experimental results demonstrate that the proposed method outperforms the compared methods, the method is also superior to all previous deep hashing methods based on triplet label.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"122 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICTAI.2019.00090","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In recent years, the emerging hashing techniques have been successful in large-scale image retrieval. Due to its strong learning ability, deep hashing has become one of the most promising solutions and achieved good results in practice. However, existing deep hashing methods had some limitations, for example, most methods consider only one kind of supervised loss, which leads to insufficient utilization of supervised information. To address this issue, we proposed a Triplet Deep Hashing method with Joint supervised Loss based on convolution neural network (JLTDH) in this work. The proposed JLTDH method combine triplet likelihood loss and linear classification loss, moreover, the triplet supervised label is adopted, which contains richer supervised information than that of pointwise and pairwise label. At the same time, in order to overcome the cubic increase in the number of triplets and make triplet training more effective, we adopt a novel triplet selection method. The whole process is divided into two stages, in the first stage, taking the triplets generated by the triplet selection method as the input of CNN, the three CNNs with shared weights are used for image feature learning, the last layer of the network outputs a preliminary hash code. In the second stage, relying on the hash code of the first stage and the joint loss function, the neural network model is further optimized so that the generated hash code has higher query precision. We perform extensive experiments on three public benchmark datasets CIFAR-10, NUS-WIDE, and MS-COCO. Experimental results demonstrate that the proposed method outperforms the compared methods, the method is also superior to all previous deep hashing methods based on triplet label.
基于联合监督损失的三重深度哈希快速图像检索
近年来,新兴的哈希技术在大规模图像检索中取得了成功。由于其强大的学习能力,深度哈希已成为最有前途的解决方案之一,并在实践中取得了良好的效果。然而,现有的深度哈希方法存在一定的局限性,例如大多数方法只考虑一种监督损失,导致监督信息利用不足。为了解决这一问题,本文提出了一种基于卷积神经网络(JLTDH)的联合监督损失三重深度哈希方法。提出的JLTDH方法结合了三联体似然损失和线性分类损失,并采用了三联体监督标记,比点向和两两标记包含更丰富的监督信息。同时,为了克服三元组数量的立方增加,使三元组训练更有效,我们采用了一种新颖的三元组选择方法。整个过程分为两个阶段,第一阶段,以三元组选择法生成的三元组作为CNN的输入,将三个权值共享的CNN用于图像特征学习,网络的最后一层输出一个初步的哈希码。第二阶段,依托第一阶段的哈希码和联合损失函数,进一步优化神经网络模型,使生成的哈希码具有更高的查询精度。我们在三个公共基准数据集CIFAR-10、NUS-WIDE和MS-COCO上进行了广泛的实验。实验结果表明,该方法在性能上优于其他方法,也优于以往基于三元组标签的深度哈希方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信