跨模态检索的语义一致深度量化

Liya Ma, N. Zhang, Kuang-I Shu, Xitao Zou
{"title":"跨模态检索的语义一致深度量化","authors":"Liya Ma, N. Zhang, Kuang-I Shu, Xitao Zou","doi":"10.1109/ICICIP53388.2021.9642180","DOIUrl":null,"url":null,"abstract":"With making up for the deficiency of the constraint representation capability of hashing codes for high-dimensional data, the quantization method has been found to generally perform better in cross-modal similarity retrieval research. However, in current quantization approaches, the codebook, as the most critical basis for quantization, is still in a passive status and detached from the learning framework. To improve the initiative of codebook, we propose a semantic-consistent deep quantization (SCDQ), which is the first scheme to integrate quantization into deep network learning in an end-to-end fashion. Specifically, two classifiers following the deep representation learning networks are formulated to produce the class-wise abstract patterns with the help of label alignment. Meanwhile, our approach learns a collaborative codebook for both modalities, which embeds bimodality semantic consistent information in codewords and bridges the relationship between the patterns in classifiers and codewords in codebook. By designing a novel algorithm architecture and codebook update strategy, SCDQ enables effective and efficient cross-modal retrieval in an asymmetric way. Extensive experiments on two benchmark datasets demonstrate that SCDQ yields optimal cross-modal retrieval performance and outperforms several state of-the-art cross-modal retrieval methods.","PeriodicalId":435799,"journal":{"name":"2021 11th International Conference on Intelligent Control and Information Processing (ICICIP)","volume":"154 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Semantic-Consistent Deep Quantization for Cross-modal Retrieval\",\"authors\":\"Liya Ma, N. Zhang, Kuang-I Shu, Xitao Zou\",\"doi\":\"10.1109/ICICIP53388.2021.9642180\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With making up for the deficiency of the constraint representation capability of hashing codes for high-dimensional data, the quantization method has been found to generally perform better in cross-modal similarity retrieval research. However, in current quantization approaches, the codebook, as the most critical basis for quantization, is still in a passive status and detached from the learning framework. To improve the initiative of codebook, we propose a semantic-consistent deep quantization (SCDQ), which is the first scheme to integrate quantization into deep network learning in an end-to-end fashion. Specifically, two classifiers following the deep representation learning networks are formulated to produce the class-wise abstract patterns with the help of label alignment. Meanwhile, our approach learns a collaborative codebook for both modalities, which embeds bimodality semantic consistent information in codewords and bridges the relationship between the patterns in classifiers and codewords in codebook. By designing a novel algorithm architecture and codebook update strategy, SCDQ enables effective and efficient cross-modal retrieval in an asymmetric way. Extensive experiments on two benchmark datasets demonstrate that SCDQ yields optimal cross-modal retrieval performance and outperforms several state of-the-art cross-modal retrieval methods.\",\"PeriodicalId\":435799,\"journal\":{\"name\":\"2021 11th International Conference on Intelligent Control and Information Processing (ICICIP)\",\"volume\":\"154 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 11th International Conference on Intelligent Control and Information Processing (ICICIP)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICICIP53388.2021.9642180\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 11th International Conference on Intelligent Control and Information Processing (ICICIP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICICIP53388.2021.9642180","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

量化方法弥补了哈希码对高维数据约束表示能力的不足,在跨模态相似性检索研究中普遍表现较好。然而,在目前的量化方法中,码本作为量化最关键的基础,仍然处于被动状态,脱离了学习框架。为了提高码本的主动性,我们提出了语义一致深度量化(SCDQ),这是第一个将量化以端到端方式集成到深度网络学习中的方案。具体来说,两个分类器遵循深度表示学习网络,在标签对齐的帮助下产生类智能抽象模式。同时,我们的方法学习了一个两模态的协作码本,它在码字中嵌入了双模态语义一致信息,并在分类器模式和码字之间架起了桥梁。通过设计一种新颖的算法架构和码本更新策略,SCDQ能够以不对称的方式实现高效的跨模态检索。在两个基准数据集上的大量实验表明,SCDQ产生了最佳的跨模态检索性能,并且优于几种最先进的跨模态检索方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Semantic-Consistent Deep Quantization for Cross-modal Retrieval
With making up for the deficiency of the constraint representation capability of hashing codes for high-dimensional data, the quantization method has been found to generally perform better in cross-modal similarity retrieval research. However, in current quantization approaches, the codebook, as the most critical basis for quantization, is still in a passive status and detached from the learning framework. To improve the initiative of codebook, we propose a semantic-consistent deep quantization (SCDQ), which is the first scheme to integrate quantization into deep network learning in an end-to-end fashion. Specifically, two classifiers following the deep representation learning networks are formulated to produce the class-wise abstract patterns with the help of label alignment. Meanwhile, our approach learns a collaborative codebook for both modalities, which embeds bimodality semantic consistent information in codewords and bridges the relationship between the patterns in classifiers and codewords in codebook. By designing a novel algorithm architecture and codebook update strategy, SCDQ enables effective and efficient cross-modal retrieval in an asymmetric way. Extensive experiments on two benchmark datasets demonstrate that SCDQ yields optimal cross-modal retrieval performance and outperforms several state of-the-art cross-modal retrieval methods.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信