{"title":"Ensemble Learning using Vision Transformer and Convolutional Networks for Person Re-ID","authors":"A. Gupta, Neil Gautam, D. Vishwakarma","doi":"10.1109/ICCMC53470.2022.9753761","DOIUrl":null,"url":null,"abstract":"Person Re-Identification is the process of recognizing a targeted individual across multiple views at different times, in different and challenging real-life diverse settings. It remains a conundrum due to the significant amount of intra-class variation present in same individual caught across different cameras. Most of the existing models require a large amount of data for training, as a result of which they do not generalize well on small datasets and hence decreases the robustness of the identification process. To reduce this variance, this paper introduces an end-to-end triple stream ensemble model making minimal changes in the Vision Transformer, Resnet50 and Densenet121 architectures respectively. Our model performs well on the Market1501 dataset achieving an accuracy of 90.05% and 80.45% on the Duke MTMC ReID dataset.","PeriodicalId":345346,"journal":{"name":"2022 6th International Conference on Computing Methodologies and Communication (ICCMC)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 6th International Conference on Computing Methodologies and Communication (ICCMC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCMC53470.2022.9753761","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Person Re-Identification is the process of recognizing a targeted individual across multiple views at different times, in different and challenging real-life diverse settings. It remains a conundrum due to the significant amount of intra-class variation present in same individual caught across different cameras. Most of the existing models require a large amount of data for training, as a result of which they do not generalize well on small datasets and hence decreases the robustness of the identification process. To reduce this variance, this paper introduces an end-to-end triple stream ensemble model making minimal changes in the Vision Transformer, Resnet50 and Densenet121 architectures respectively. Our model performs well on the Market1501 dataset achieving an accuracy of 90.05% and 80.45% on the Duke MTMC ReID dataset.