Expanding and Refining Hybrid Compressors for Efficient Object Re-Identification

Yi Xie;Hanxiao Wu;Jianqing Zhu;Huanqiang Zeng;Jing Zhang
{"title":"Expanding and Refining Hybrid Compressors for Efficient Object Re-Identification","authors":"Yi Xie;Hanxiao Wu;Jianqing Zhu;Huanqiang Zeng;Jing Zhang","doi":"10.1109/TIP.2024.3410684","DOIUrl":null,"url":null,"abstract":"Recent object re-identification (Re-ID) methods gain high efficiency via lightweight student models trained by knowledge distillation (KD). However, the huge architectural difference between lightweight students and heavy teachers causes students to have difficulties in receiving and understanding teachers’ knowledge, thus losing certain accuracy. To this end, we propose a refiner-expander-refiner (RER) structure to enlarge a student’s representational capacity and prune the student’s complexity. The expander is a multi-branch convolutional layer to expand the student’s representational capacity to understand a teacher’s knowledge comprehensively, which does not require any feature-dimensional adapter to avoid knowledge distortions. The two refiners are \n<inline-formula> <tex-math>$1\\times 1$ </tex-math></inline-formula>\n convolutional layers to prune the input and output channels of the expander. In addition, in order to alleviate the competition accuracy-related and pruning-related gradients, we design a common consensus gradient resetting (CCGR) method, which discards unimportant channels according to the intersection of each sample’s unimportant channel judgment. Finally, the trained RER can be simplified into a slim convolutional layer via re-parameterization to speed up inference. As a result, we propose an expanding and refining hybrid compressing (ERHC) method. Extensive experiments show that our ERHC has superior inference speed and accuracy, e.g., on the VeRi-776 dataset, given the ResNet101 as a teacher, ERHC saves 75.33% model parameters (MP) and 74.29% floating-point of operations (FLOPs) without sacrificing accuracy.","PeriodicalId":94032,"journal":{"name":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-06-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on image processing : a publication of the IEEE Signal Processing Society","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10555505/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Recent object re-identification (Re-ID) methods gain high efficiency via lightweight student models trained by knowledge distillation (KD). However, the huge architectural difference between lightweight students and heavy teachers causes students to have difficulties in receiving and understanding teachers’ knowledge, thus losing certain accuracy. To this end, we propose a refiner-expander-refiner (RER) structure to enlarge a student’s representational capacity and prune the student’s complexity. The expander is a multi-branch convolutional layer to expand the student’s representational capacity to understand a teacher’s knowledge comprehensively, which does not require any feature-dimensional adapter to avoid knowledge distortions. The two refiners are $1\times 1$ convolutional layers to prune the input and output channels of the expander. In addition, in order to alleviate the competition accuracy-related and pruning-related gradients, we design a common consensus gradient resetting (CCGR) method, which discards unimportant channels according to the intersection of each sample’s unimportant channel judgment. Finally, the trained RER can be simplified into a slim convolutional layer via re-parameterization to speed up inference. As a result, we propose an expanding and refining hybrid compressing (ERHC) method. Extensive experiments show that our ERHC has superior inference speed and accuracy, e.g., on the VeRi-776 dataset, given the ResNet101 as a teacher, ERHC saves 75.33% model parameters (MP) and 74.29% floating-point of operations (FLOPs) without sacrificing accuracy.
扩展和改进混合压缩机,实现高效的对象再识别。
最近的物体再识别(Re-ID)方法通过知识蒸馏(KD)训练的轻量级学生模型获得了很高的效率。然而,轻量级学生与重度教师之间巨大的结构差异导致学生难以接收和理解教师的知识,从而失去了一定的准确性。为此,我们提出了一种提炼器-扩展器-提炼器(RER)结构,以扩大学生的表征能力,删减学生的复杂性。扩展器是一个多分支卷积层,用于扩展学生的表征能力,以全面理解教师的知识,它不需要任何特征维度的适配器,以避免知识失真。两个细化器为 1 × 1 卷积层,用于修剪扩展器的输入和输出通道。此外,为了缓解竞争精度相关梯度和修剪相关梯度,我们设计了一种共同共识梯度重置(CCGR)方法,根据每个样本不重要通道判断的交集来丢弃不重要通道。最后,经过训练的 RER 可以通过重新参数化简化为纤细的卷积层,从而加快推理速度。因此,我们提出了一种扩展和细化混合压缩(ERHC)方法。广泛的实验表明,我们的ERHC推理速度和准确性都很出色,例如,在VeRi-776数据集上,以ResNet101为老师,ERHC在不牺牲准确性的情况下节省了75.33%的模型参数(MP)和74.29%的浮点运算(FLOPs)。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信