Jianqun Zhou;Yang Li;Hongmao Qin;Pengwen Dai;Zilong Zhao;Manjiang Hu
{"title":"利用 MFA-CycleGAN 生成声纳图像,增强 AUV 的水下物体探测能力","authors":"Jianqun Zhou;Yang Li;Hongmao Qin;Pengwen Dai;Zilong Zhao;Manjiang Hu","doi":"10.1109/JOE.2024.3350746","DOIUrl":null,"url":null,"abstract":"Acquiring large amounts of high-quality real sonar data for object detection of autonomous underwater vehicles (AUVs) is challenging. Synthetic data can be an alternative, but it is hard to generate diverse data using traditional generative models when real data are limited. This study proposes a novel style transfer method, i.e., the multigranular feature alignment cycle-consistent generative adversarial network (CycleGAN), to generate sonar images leveraging remote sensing images, which can alleviate the dependence on real sonar data. Specifically, we add a spatial attention-based feature aggregation module to preserve unique features by attending to instance parts of an image. A pair of cross-domain discriminators are designed to guide generators to produce images that capture sonar styles. We also introduce a novel cycle consistency loss based on the discrete cosine transform of images, which better utilizes features that are evident in the frequency domain. Extensive experimental results show that the generated sonar images have better quality than CycleGAN, with improvements of 15.2% in IS, 56.9% in FID, 42.6% in KID, and 7.6% in learned perceptual image patch similarity, respectively. Moreover, after expanding the real sonar dataset with generated data, the average accuracy of the object detector, e.g., YOLOv6, has increased by more than 48.7%, indicating the effectiveness of the generated sonar data by our method.","PeriodicalId":13191,"journal":{"name":"IEEE Journal of Oceanic Engineering","volume":"49 3","pages":"905-919"},"PeriodicalIF":3.8000,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Sonar Image Generation by MFA-CycleGAN for Boosting Underwater Object Detection of AUVs\",\"authors\":\"Jianqun Zhou;Yang Li;Hongmao Qin;Pengwen Dai;Zilong Zhao;Manjiang Hu\",\"doi\":\"10.1109/JOE.2024.3350746\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Acquiring large amounts of high-quality real sonar data for object detection of autonomous underwater vehicles (AUVs) is challenging. Synthetic data can be an alternative, but it is hard to generate diverse data using traditional generative models when real data are limited. This study proposes a novel style transfer method, i.e., the multigranular feature alignment cycle-consistent generative adversarial network (CycleGAN), to generate sonar images leveraging remote sensing images, which can alleviate the dependence on real sonar data. Specifically, we add a spatial attention-based feature aggregation module to preserve unique features by attending to instance parts of an image. A pair of cross-domain discriminators are designed to guide generators to produce images that capture sonar styles. We also introduce a novel cycle consistency loss based on the discrete cosine transform of images, which better utilizes features that are evident in the frequency domain. Extensive experimental results show that the generated sonar images have better quality than CycleGAN, with improvements of 15.2% in IS, 56.9% in FID, 42.6% in KID, and 7.6% in learned perceptual image patch similarity, respectively. Moreover, after expanding the real sonar dataset with generated data, the average accuracy of the object detector, e.g., YOLOv6, has increased by more than 48.7%, indicating the effectiveness of the generated sonar data by our method.\",\"PeriodicalId\":13191,\"journal\":{\"name\":\"IEEE Journal of Oceanic Engineering\",\"volume\":\"49 3\",\"pages\":\"905-919\"},\"PeriodicalIF\":3.8000,\"publicationDate\":\"2024-03-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Journal of Oceanic Engineering\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10472044/\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, CIVIL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Journal of Oceanic Engineering","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10472044/","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, CIVIL","Score":null,"Total":0}
Sonar Image Generation by MFA-CycleGAN for Boosting Underwater Object Detection of AUVs
Acquiring large amounts of high-quality real sonar data for object detection of autonomous underwater vehicles (AUVs) is challenging. Synthetic data can be an alternative, but it is hard to generate diverse data using traditional generative models when real data are limited. This study proposes a novel style transfer method, i.e., the multigranular feature alignment cycle-consistent generative adversarial network (CycleGAN), to generate sonar images leveraging remote sensing images, which can alleviate the dependence on real sonar data. Specifically, we add a spatial attention-based feature aggregation module to preserve unique features by attending to instance parts of an image. A pair of cross-domain discriminators are designed to guide generators to produce images that capture sonar styles. We also introduce a novel cycle consistency loss based on the discrete cosine transform of images, which better utilizes features that are evident in the frequency domain. Extensive experimental results show that the generated sonar images have better quality than CycleGAN, with improvements of 15.2% in IS, 56.9% in FID, 42.6% in KID, and 7.6% in learned perceptual image patch similarity, respectively. Moreover, after expanding the real sonar dataset with generated data, the average accuracy of the object detector, e.g., YOLOv6, has increased by more than 48.7%, indicating the effectiveness of the generated sonar data by our method.
期刊介绍:
The IEEE Journal of Oceanic Engineering (ISSN 0364-9059) is the online-only quarterly publication of the IEEE Oceanic Engineering Society (IEEE OES). The scope of the Journal is the field of interest of the IEEE OES, which encompasses all aspects of science, engineering, and technology that address research, development, and operations pertaining to all bodies of water. This includes the creation of new capabilities and technologies from concept design through prototypes, testing, and operational systems to sense, explore, understand, develop, use, and responsibly manage natural resources.