Sonar Image Generation by MFA-CycleGAN for Boosting Underwater Object Detection of AUVs

IF 3.8 2区 工程技术 Q1 ENGINEERING, CIVIL
Jianqun Zhou;Yang Li;Hongmao Qin;Pengwen Dai;Zilong Zhao;Manjiang Hu
{"title":"Sonar Image Generation by MFA-CycleGAN for Boosting Underwater Object Detection of AUVs","authors":"Jianqun Zhou;Yang Li;Hongmao Qin;Pengwen Dai;Zilong Zhao;Manjiang Hu","doi":"10.1109/JOE.2024.3350746","DOIUrl":null,"url":null,"abstract":"Acquiring large amounts of high-quality real sonar data for object detection of autonomous underwater vehicles (AUVs) is challenging. Synthetic data can be an alternative, but it is hard to generate diverse data using traditional generative models when real data are limited. This study proposes a novel style transfer method, i.e., the multigranular feature alignment cycle-consistent generative adversarial network (CycleGAN), to generate sonar images leveraging remote sensing images, which can alleviate the dependence on real sonar data. Specifically, we add a spatial attention-based feature aggregation module to preserve unique features by attending to instance parts of an image. A pair of cross-domain discriminators are designed to guide generators to produce images that capture sonar styles. We also introduce a novel cycle consistency loss based on the discrete cosine transform of images, which better utilizes features that are evident in the frequency domain. Extensive experimental results show that the generated sonar images have better quality than CycleGAN, with improvements of 15.2% in IS, 56.9% in FID, 42.6% in KID, and 7.6% in learned perceptual image patch similarity, respectively. Moreover, after expanding the real sonar dataset with generated data, the average accuracy of the object detector, e.g., YOLOv6, has increased by more than 48.7%, indicating the effectiveness of the generated sonar data by our method.","PeriodicalId":13191,"journal":{"name":"IEEE Journal of Oceanic Engineering","volume":"49 3","pages":"905-919"},"PeriodicalIF":3.8000,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Journal of Oceanic Engineering","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10472044/","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, CIVIL","Score":null,"Total":0}
引用次数: 0

Abstract

Acquiring large amounts of high-quality real sonar data for object detection of autonomous underwater vehicles (AUVs) is challenging. Synthetic data can be an alternative, but it is hard to generate diverse data using traditional generative models when real data are limited. This study proposes a novel style transfer method, i.e., the multigranular feature alignment cycle-consistent generative adversarial network (CycleGAN), to generate sonar images leveraging remote sensing images, which can alleviate the dependence on real sonar data. Specifically, we add a spatial attention-based feature aggregation module to preserve unique features by attending to instance parts of an image. A pair of cross-domain discriminators are designed to guide generators to produce images that capture sonar styles. We also introduce a novel cycle consistency loss based on the discrete cosine transform of images, which better utilizes features that are evident in the frequency domain. Extensive experimental results show that the generated sonar images have better quality than CycleGAN, with improvements of 15.2% in IS, 56.9% in FID, 42.6% in KID, and 7.6% in learned perceptual image patch similarity, respectively. Moreover, after expanding the real sonar dataset with generated data, the average accuracy of the object detector, e.g., YOLOv6, has increased by more than 48.7%, indicating the effectiveness of the generated sonar data by our method.
利用 MFA-CycleGAN 生成声纳图像,增强 AUV 的水下物体探测能力
获取大量高质量的真实声纳数据用于自动潜航器(AUV)的物体探测是一项挑战。合成数据可以作为一种替代方法,但在真实数据有限的情况下,使用传统的生成模型很难生成多样化的数据。本研究提出了一种新颖的样式转移方法,即多粒度特征配准循环-一致性生成对抗网络(CycleGAN),利用遥感图像生成声纳图像,从而减轻对真实声纳数据的依赖。具体来说,我们添加了一个基于空间注意力的特征聚合模块,通过关注图像的实例部分来保留独特的特征。我们设计了一对跨域判别器,引导生成器生成捕捉声纳风格的图像。我们还在图像离散余弦变换的基础上引入了一种新的周期一致性损失,它能更好地利用频域中明显的特征。广泛的实验结果表明,生成的声纳图像质量优于 CycleGAN,IS、FID、KID 和学习感知图像补丁相似度分别提高了 15.2%、56.9%、42.6% 和 7.6%。此外,用生成的数据扩展真实声纳数据集后,物体检测器(如 YOLOv6)的平均准确率提高了 48.7% 以上,这表明我们的方法生成的声纳数据非常有效。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
IEEE Journal of Oceanic Engineering
IEEE Journal of Oceanic Engineering 工程技术-工程:大洋
CiteScore
9.60
自引率
12.20%
发文量
86
审稿时长
12 months
期刊介绍: The IEEE Journal of Oceanic Engineering (ISSN 0364-9059) is the online-only quarterly publication of the IEEE Oceanic Engineering Society (IEEE OES). The scope of the Journal is the field of interest of the IEEE OES, which encompasses all aspects of science, engineering, and technology that address research, development, and operations pertaining to all bodies of water. This includes the creation of new capabilities and technologies from concept design through prototypes, testing, and operational systems to sense, explore, understand, develop, use, and responsibly manage natural resources.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信