Large-Scale Fully-Unsupervised Re-Identification

Gabriel Bertocco;Fernanda Andaló;Terrance E. Boult;Anderson Rocha
{"title":"Large-Scale Fully-Unsupervised Re-Identification","authors":"Gabriel Bertocco;Fernanda Andaló;Terrance E. Boult;Anderson Rocha","doi":"10.1109/TBIOM.2024.3446964","DOIUrl":null,"url":null,"abstract":"Fully-unsupervised Person and Vehicle Re-Identification have received increasing attention due to their broad applicability in areas such as surveillance, forensics, event understanding, and smart cities, without requiring any manual annotation. However, most of the prior art has been evaluated in datasets that have just a couple thousand samples. Such small-data setups often allow the use of costly techniques in terms of time and memory footprints, such as Re-Ranking, to improve clustering results. Moreover, some previous work even pre-selects the best clustering hyper-parameters for each dataset, which is unrealistic in a large-scale fully-unsupervised scenario. In this context, this work tackles a more realistic scenario and proposes two strategies to learn from large-scale unlabeled data. The first strategy performs a local neighborhood sampling to reduce the dataset size in each iteration without violating neighborhood relationships. A second strategy leverages a novel Re-Ranking technique, which has a lower time upper bound complexity and reduces the memory complexity from <inline-formula> <tex-math>$\\mathcal {O}(n^{2})$ </tex-math></inline-formula> to <inline-formula> <tex-math>$\\mathcal {O}(kn)$ </tex-math></inline-formula> with <inline-formula> <tex-math>$k \\ll n$ </tex-math></inline-formula>. To avoid the need for pre-selection of specific hyper-parameter values for the clustering algorithm, we also present a novel scheduling algorithm that adjusts the density parameter during training, to leverage the diversity of samples and keep the learning robust to noisy labeling. Finally, due to the complementary knowledge learned by different models in an ensemble, we also introduce a co-training strategy that relies upon the permutation of predicted pseudo-labels, among the backbones, with no need for any hyper-parameters or weighting optimization. The proposed methodology outperforms the state-of-the-art methods in well-known benchmarks and in the challenging large-scale Veri-Wild dataset, with a faster and memory-efficient Re-Ranking strategy, and a large-scale, noisy-robust, and ensemble-based learning approach.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"7 2","pages":"156-169"},"PeriodicalIF":0.0000,"publicationDate":"2024-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on biometrics, behavior, and identity science","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10643149/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Fully-unsupervised Person and Vehicle Re-Identification have received increasing attention due to their broad applicability in areas such as surveillance, forensics, event understanding, and smart cities, without requiring any manual annotation. However, most of the prior art has been evaluated in datasets that have just a couple thousand samples. Such small-data setups often allow the use of costly techniques in terms of time and memory footprints, such as Re-Ranking, to improve clustering results. Moreover, some previous work even pre-selects the best clustering hyper-parameters for each dataset, which is unrealistic in a large-scale fully-unsupervised scenario. In this context, this work tackles a more realistic scenario and proposes two strategies to learn from large-scale unlabeled data. The first strategy performs a local neighborhood sampling to reduce the dataset size in each iteration without violating neighborhood relationships. A second strategy leverages a novel Re-Ranking technique, which has a lower time upper bound complexity and reduces the memory complexity from $\mathcal {O}(n^{2})$ to $\mathcal {O}(kn)$ with $k \ll n$ . To avoid the need for pre-selection of specific hyper-parameter values for the clustering algorithm, we also present a novel scheduling algorithm that adjusts the density parameter during training, to leverage the diversity of samples and keep the learning robust to noisy labeling. Finally, due to the complementary knowledge learned by different models in an ensemble, we also introduce a co-training strategy that relies upon the permutation of predicted pseudo-labels, among the backbones, with no need for any hyper-parameters or weighting optimization. The proposed methodology outperforms the state-of-the-art methods in well-known benchmarks and in the challenging large-scale Veri-Wild dataset, with a faster and memory-efficient Re-Ranking strategy, and a large-scale, noisy-robust, and ensemble-based learning approach.
大规模的完全无监督再识别
完全无监督的人员和车辆再识别由于其在监视,取证,事件理解和智能城市等领域的广泛适用性而受到越来越多的关注,而无需任何手动注释。然而,大多数现有技术都是在只有几千个样本的数据集中进行评估的。这种小数据设置通常允许使用在时间和内存占用方面代价高昂的技术,例如重新排序,以改善聚类结果。此外,之前的一些工作甚至为每个数据集预先选择了最佳聚类超参数,这在大规模的完全无监督场景中是不现实的。在此背景下,本工作解决了一个更现实的场景,并提出了两种从大规模未标记数据中学习的策略。第一种策略是在不破坏邻域关系的情况下执行局部邻域采样,以减少每次迭代中的数据集大小。第二种策略利用了一种新颖的重新排序技术,该技术具有较低的时间上限复杂度,并将内存复杂度从$\mathcal {O}(n^{2})$降低到$\mathcal {O}(kn)$和$k \ll n$。为了避免聚类算法需要预先选择特定的超参数值,我们还提出了一种新的调度算法,该算法在训练过程中调整密度参数,以利用样本的多样性并保持学习对噪声标记的鲁棒性。最后,由于集成中不同模型学习到的知识是互补的,我们还引入了一种协同训练策略,该策略依赖于预测伪标签的排列,在主干中,不需要任何超参数或加权优化。所提出的方法在众所周知的基准测试和具有挑战性的大规模Veri-Wild数据集中优于最先进的方法,具有更快和内存效率的重新排名策略,以及大规模,噪声鲁棒性和基于集成的学习方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
10.90
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信