{"title":"暗光照条件下合成图像到实像的无监督转换方法","authors":"Zhixiong Wang, Wei Chen, Zhenyu Fang, Haoyang Zhang, Liang Xie, Ye Yan, Erwei Yin","doi":"10.1117/12.3004756","DOIUrl":null,"url":null,"abstract":"Most of the existing hand data collection procedures rely on complex and bulky image acquisition systems, in fact, the data directly acquired by the acquisition systems do not have a realistic context, and the neural network models trained from these data perform not well in real-world situations, especially when people are wearing gloves in scenes with changing illumination. However, to improve the consistency of the image the traditional image fusion methods sacrifice the authenticity of the background and the supervised methods are limited by the data acquisition system which cannot generate high quality fusion image. Therefore, based on an unsupervised generative adversarial network, this paper introduced a generator based on a global attention module and proposed a histogram similarity-based image selection module to filter the input images. Our goal is to decrease the difficulty of the image migration task for the network. The model had been trained using a self-built dataset consisting of composited and real data, including images of hands wearing IMU data gloves and photographs taken in real life with similar backgrounds. Extensive experiments on image segmentation tasks demonstrated the effectiveness of the proposed model, which obtained a precision of 0.88 and a recall of 0.86 with a mean NIMA value of 4.27 compared to other state-of-the-art methods such as UEGAN and DoveNet.","PeriodicalId":143265,"journal":{"name":"6th International Conference on Advanced Electronic Materials, Computers and Software Engineering (AEMCSE 2023)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An unsupervised transfer method for composited image to real image under dark light conditions\",\"authors\":\"Zhixiong Wang, Wei Chen, Zhenyu Fang, Haoyang Zhang, Liang Xie, Ye Yan, Erwei Yin\",\"doi\":\"10.1117/12.3004756\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Most of the existing hand data collection procedures rely on complex and bulky image acquisition systems, in fact, the data directly acquired by the acquisition systems do not have a realistic context, and the neural network models trained from these data perform not well in real-world situations, especially when people are wearing gloves in scenes with changing illumination. However, to improve the consistency of the image the traditional image fusion methods sacrifice the authenticity of the background and the supervised methods are limited by the data acquisition system which cannot generate high quality fusion image. Therefore, based on an unsupervised generative adversarial network, this paper introduced a generator based on a global attention module and proposed a histogram similarity-based image selection module to filter the input images. Our goal is to decrease the difficulty of the image migration task for the network. The model had been trained using a self-built dataset consisting of composited and real data, including images of hands wearing IMU data gloves and photographs taken in real life with similar backgrounds. Extensive experiments on image segmentation tasks demonstrated the effectiveness of the proposed model, which obtained a precision of 0.88 and a recall of 0.86 with a mean NIMA value of 4.27 compared to other state-of-the-art methods such as UEGAN and DoveNet.\",\"PeriodicalId\":143265,\"journal\":{\"name\":\"6th International Conference on Advanced Electronic Materials, Computers and Software Engineering (AEMCSE 2023)\",\"volume\":\"22 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-08-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"6th International Conference on Advanced Electronic Materials, Computers and Software Engineering (AEMCSE 2023)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1117/12.3004756\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"6th International Conference on Advanced Electronic Materials, Computers and Software Engineering (AEMCSE 2023)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.3004756","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
An unsupervised transfer method for composited image to real image under dark light conditions
Most of the existing hand data collection procedures rely on complex and bulky image acquisition systems, in fact, the data directly acquired by the acquisition systems do not have a realistic context, and the neural network models trained from these data perform not well in real-world situations, especially when people are wearing gloves in scenes with changing illumination. However, to improve the consistency of the image the traditional image fusion methods sacrifice the authenticity of the background and the supervised methods are limited by the data acquisition system which cannot generate high quality fusion image. Therefore, based on an unsupervised generative adversarial network, this paper introduced a generator based on a global attention module and proposed a histogram similarity-based image selection module to filter the input images. Our goal is to decrease the difficulty of the image migration task for the network. The model had been trained using a self-built dataset consisting of composited and real data, including images of hands wearing IMU data gloves and photographs taken in real life with similar backgrounds. Extensive experiments on image segmentation tasks demonstrated the effectiveness of the proposed model, which obtained a precision of 0.88 and a recall of 0.86 with a mean NIMA value of 4.27 compared to other state-of-the-art methods such as UEGAN and DoveNet.