{"title":"属性-图像人再识别的双模态元度量学习","authors":"Rongxian Xu, Fei Shen, Hanxiao Wu, Jianqing Zhu, Huanqiang Zeng","doi":"10.1109/ICNSC52481.2021.9702261","DOIUrl":null,"url":null,"abstract":"Attribute-image person re-identification (AIPR) aiming to retrieve persons from massive images via an attribute query is a meaningful but challenging cross-modal retrieval task. Because there is a huge modal difference between person images and attributes, e.g., on the image modal one subject usually contains of varying instances, but on the attribute modal, one subject only contains an explicit instance. Unlike most existing AIPR methods focusing on shrinking feature differences crossing modals, we propose a dual modal meta metric learning (DM3L) method for AIPR in this paper. Specifically, in each episode, we sample a subset as a new task and split the training data into a single-modal support set of person images and a dual modal query set consisting of both person images and attributes. Based on the single-modal support set and the dual modal query set, our DM3L learns not only attribute-image cross-modal metrics but also learns image-image intra-modal metrics. Therefore, our DM3L method encourages data on both attribute and image modalities are discriminate to improve AIPR. Experiments show that our DM3L outperforms state-of-the-art approaches on Market-1501 Attribute and PETA datasets.","PeriodicalId":129062,"journal":{"name":"2021 IEEE International Conference on Networking, Sensing and Control (ICNSC)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Dual Modal Meta Metric Learning for Attribute-Image Person Re-identification\",\"authors\":\"Rongxian Xu, Fei Shen, Hanxiao Wu, Jianqing Zhu, Huanqiang Zeng\",\"doi\":\"10.1109/ICNSC52481.2021.9702261\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Attribute-image person re-identification (AIPR) aiming to retrieve persons from massive images via an attribute query is a meaningful but challenging cross-modal retrieval task. Because there is a huge modal difference between person images and attributes, e.g., on the image modal one subject usually contains of varying instances, but on the attribute modal, one subject only contains an explicit instance. Unlike most existing AIPR methods focusing on shrinking feature differences crossing modals, we propose a dual modal meta metric learning (DM3L) method for AIPR in this paper. Specifically, in each episode, we sample a subset as a new task and split the training data into a single-modal support set of person images and a dual modal query set consisting of both person images and attributes. Based on the single-modal support set and the dual modal query set, our DM3L learns not only attribute-image cross-modal metrics but also learns image-image intra-modal metrics. Therefore, our DM3L method encourages data on both attribute and image modalities are discriminate to improve AIPR. Experiments show that our DM3L outperforms state-of-the-art approaches on Market-1501 Attribute and PETA datasets.\",\"PeriodicalId\":129062,\"journal\":{\"name\":\"2021 IEEE International Conference on Networking, Sensing and Control (ICNSC)\",\"volume\":\"2 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE International Conference on Networking, Sensing and Control (ICNSC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICNSC52481.2021.9702261\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Networking, Sensing and Control (ICNSC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICNSC52481.2021.9702261","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Dual Modal Meta Metric Learning for Attribute-Image Person Re-identification
Attribute-image person re-identification (AIPR) aiming to retrieve persons from massive images via an attribute query is a meaningful but challenging cross-modal retrieval task. Because there is a huge modal difference between person images and attributes, e.g., on the image modal one subject usually contains of varying instances, but on the attribute modal, one subject only contains an explicit instance. Unlike most existing AIPR methods focusing on shrinking feature differences crossing modals, we propose a dual modal meta metric learning (DM3L) method for AIPR in this paper. Specifically, in each episode, we sample a subset as a new task and split the training data into a single-modal support set of person images and a dual modal query set consisting of both person images and attributes. Based on the single-modal support set and the dual modal query set, our DM3L learns not only attribute-image cross-modal metrics but also learns image-image intra-modal metrics. Therefore, our DM3L method encourages data on both attribute and image modalities are discriminate to improve AIPR. Experiments show that our DM3L outperforms state-of-the-art approaches on Market-1501 Attribute and PETA datasets.