{"title":"一种用于人再识别数据增强的改进CycleGAN","authors":"Zhenzhen Yang , Jing Shao , Yongpeng Yang","doi":"10.1016/j.bdr.2023.100409","DOIUrl":null,"url":null,"abstract":"<div><p>Person re-identification (ReID) has attracted more and more attention, which is to retrieve interested persons across multiple non-overlapping cameras. Matching the same person between different camera styles has always been an enormous challenge. In the existing work, cross-camera styles images generated by the cycle-consistent generative adversarial network<span> (CycleGAN) only transfer the camera resolution and ambient lighting. The generated images produce considerable redundancy and inappropriate pictures at the same time. Although the data is added to prevent over-fitting, it also makes significant noise, so the accuracy is not significantly improved. In this paper, an improved CycleGAN is proposed to generate images for achieving improved data augmentation. The transfer of pedestrian posture is added at the same time as transferring the image style. It not only increases the diversity of pedestrian posture but also reduces the domain gap caused by the style change between cameras. Besides, through the multi-pseudo regularized label (MpRL), the generated images are assigned virtual tags dynamically in training. Through many experimental evaluations, we have achieved a very high identification accuracy on Market-1501, DukeMTMC-reID, and CUHK03-NP datasets. On the three datasets, the quantitative results of mAP are 96.20%, 93.72%, and 86.65%, and the quantitative results of rank-1 are 98.27%, 95.37%, and 90.71%, respectively. The experimental results fully show the superiority of our proposed method.</span></p></div>","PeriodicalId":56017,"journal":{"name":"Big Data Research","volume":"34 ","pages":"Article 100409"},"PeriodicalIF":3.5000,"publicationDate":"2023-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Improved CycleGAN for Data Augmentation in Person Re-Identification\",\"authors\":\"Zhenzhen Yang , Jing Shao , Yongpeng Yang\",\"doi\":\"10.1016/j.bdr.2023.100409\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Person re-identification (ReID) has attracted more and more attention, which is to retrieve interested persons across multiple non-overlapping cameras. Matching the same person between different camera styles has always been an enormous challenge. In the existing work, cross-camera styles images generated by the cycle-consistent generative adversarial network<span> (CycleGAN) only transfer the camera resolution and ambient lighting. The generated images produce considerable redundancy and inappropriate pictures at the same time. Although the data is added to prevent over-fitting, it also makes significant noise, so the accuracy is not significantly improved. In this paper, an improved CycleGAN is proposed to generate images for achieving improved data augmentation. The transfer of pedestrian posture is added at the same time as transferring the image style. It not only increases the diversity of pedestrian posture but also reduces the domain gap caused by the style change between cameras. Besides, through the multi-pseudo regularized label (MpRL), the generated images are assigned virtual tags dynamically in training. Through many experimental evaluations, we have achieved a very high identification accuracy on Market-1501, DukeMTMC-reID, and CUHK03-NP datasets. On the three datasets, the quantitative results of mAP are 96.20%, 93.72%, and 86.65%, and the quantitative results of rank-1 are 98.27%, 95.37%, and 90.71%, respectively. The experimental results fully show the superiority of our proposed method.</span></p></div>\",\"PeriodicalId\":56017,\"journal\":{\"name\":\"Big Data Research\",\"volume\":\"34 \",\"pages\":\"Article 100409\"},\"PeriodicalIF\":3.5000,\"publicationDate\":\"2023-09-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Big Data Research\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2214579623000424\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Big Data Research","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2214579623000424","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
An Improved CycleGAN for Data Augmentation in Person Re-Identification
Person re-identification (ReID) has attracted more and more attention, which is to retrieve interested persons across multiple non-overlapping cameras. Matching the same person between different camera styles has always been an enormous challenge. In the existing work, cross-camera styles images generated by the cycle-consistent generative adversarial network (CycleGAN) only transfer the camera resolution and ambient lighting. The generated images produce considerable redundancy and inappropriate pictures at the same time. Although the data is added to prevent over-fitting, it also makes significant noise, so the accuracy is not significantly improved. In this paper, an improved CycleGAN is proposed to generate images for achieving improved data augmentation. The transfer of pedestrian posture is added at the same time as transferring the image style. It not only increases the diversity of pedestrian posture but also reduces the domain gap caused by the style change between cameras. Besides, through the multi-pseudo regularized label (MpRL), the generated images are assigned virtual tags dynamically in training. Through many experimental evaluations, we have achieved a very high identification accuracy on Market-1501, DukeMTMC-reID, and CUHK03-NP datasets. On the three datasets, the quantitative results of mAP are 96.20%, 93.72%, and 86.65%, and the quantitative results of rank-1 are 98.27%, 95.37%, and 90.71%, respectively. The experimental results fully show the superiority of our proposed method.
期刊介绍:
The journal aims to promote and communicate advances in big data research by providing a fast and high quality forum for researchers, practitioners and policy makers from the very many different communities working on, and with, this topic.
The journal will accept papers on foundational aspects in dealing with big data, as well as papers on specific Platforms and Technologies used to deal with big data. To promote Data Science and interdisciplinary collaboration between fields, and to showcase the benefits of data driven research, papers demonstrating applications of big data in domains as diverse as Geoscience, Social Web, Finance, e-Commerce, Health Care, Environment and Climate, Physics and Astronomy, Chemistry, life sciences and drug discovery, digital libraries and scientific publications, security and government will also be considered. Occasionally the journal may publish whitepapers on policies, standards and best practices.