SynFAGnet:用于生成真实火灾图像的全自动生成网络

IF 2.3 3区 工程技术 Q2 ENGINEERING, MULTIDISCIPLINARY
Quoc Dung Nguyen, Ngoc Dau Mai, Van Huan Nguyen, Vijay Kakani, Hakil Kim
{"title":"SynFAGnet:用于生成真实火灾图像的全自动生成网络","authors":"Quoc Dung Nguyen,&nbsp;Ngoc Dau Mai,&nbsp;Van Huan Nguyen,&nbsp;Vijay Kakani,&nbsp;Hakil Kim","doi":"10.1007/s10694-023-01540-2","DOIUrl":null,"url":null,"abstract":"<div><p>This paper proposes a fully automated generative network (“SynFAGnet”) for automatically creating a realistic-looking synthetic fire image. SynFAGnet is used as a data augmentation technique to create diverse data for training models, thereby solving problems related to real data acquisition and data imbalances. SynFAGnet comprises two main parts: an object-scene placement net (OSPNet) and a local–global context-based generative adversarial network (LGC-GAN). The OSPNet identifies suitable positions and scales for fires corresponding to the background scene. The LGC-GAN enhances the realistic appearance of synthetic fire images created by a given fire object-background scene pair by assembling effects such as halos and reflections in the surrounding area in the background scene. A comparative analysis shows that SynFAGnet achieves better outcomes than previous studies for both the Fréchet inception distance and learned perceptual image patch similarity evaluation metrics (values of 17.232 and 0.077, respectively). In addition, SynFAGnet is verified as a practically applicable data augmentation technique for training datasets, as it improves the detection and instance segmentation performance.</p></div>","PeriodicalId":558,"journal":{"name":"Fire Technology","volume":null,"pages":null},"PeriodicalIF":2.3000,"publicationDate":"2024-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10694-023-01540-2.pdf","citationCount":"0","resultStr":"{\"title\":\"SynFAGnet: A Fully Automated Generative Network for Realistic Fire Image Generation\",\"authors\":\"Quoc Dung Nguyen,&nbsp;Ngoc Dau Mai,&nbsp;Van Huan Nguyen,&nbsp;Vijay Kakani,&nbsp;Hakil Kim\",\"doi\":\"10.1007/s10694-023-01540-2\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>This paper proposes a fully automated generative network (“SynFAGnet”) for automatically creating a realistic-looking synthetic fire image. SynFAGnet is used as a data augmentation technique to create diverse data for training models, thereby solving problems related to real data acquisition and data imbalances. SynFAGnet comprises two main parts: an object-scene placement net (OSPNet) and a local–global context-based generative adversarial network (LGC-GAN). The OSPNet identifies suitable positions and scales for fires corresponding to the background scene. The LGC-GAN enhances the realistic appearance of synthetic fire images created by a given fire object-background scene pair by assembling effects such as halos and reflections in the surrounding area in the background scene. A comparative analysis shows that SynFAGnet achieves better outcomes than previous studies for both the Fréchet inception distance and learned perceptual image patch similarity evaluation metrics (values of 17.232 and 0.077, respectively). In addition, SynFAGnet is verified as a practically applicable data augmentation technique for training datasets, as it improves the detection and instance segmentation performance.</p></div>\",\"PeriodicalId\":558,\"journal\":{\"name\":\"Fire Technology\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":2.3000,\"publicationDate\":\"2024-02-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://link.springer.com/content/pdf/10.1007/s10694-023-01540-2.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Fire Technology\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://link.springer.com/article/10.1007/s10694-023-01540-2\",\"RegionNum\":3,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENGINEERING, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Fire Technology","FirstCategoryId":"5","ListUrlMain":"https://link.springer.com/article/10.1007/s10694-023-01540-2","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

摘要

本文提出了一种全自动生成网络("SynFAGnet"),用于自动创建逼真的合成火灾图像。SynFAGnet 用作数据增强技术,为训练模型创建多样化数据,从而解决与真实数据采集和数据不平衡相关的问题。SynFAGnet 由两个主要部分组成:对象-场景放置网(OSPNet)和基于局部-全局上下文的生成对抗网络(LGC-GAN)。OSPNet 可根据背景场景确定火灾的合适位置和规模。LGC-GAN 通过组合背景场景中周围区域的光晕和反射等效果,增强了由给定火灾对象-背景场景对创建的合成火灾图像的逼真度。对比分析表明,SynFAGnet 在弗雷谢特插入距离和学习感知图像补丁相似性评价指标(值分别为 17.232 和 0.077)方面都取得了比以往研究更好的结果。此外,由于 SynFAGnet 提高了检测和实例分割性能,因此被证实是一种实际适用的训练数据集数据增强技术。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

SynFAGnet: A Fully Automated Generative Network for Realistic Fire Image Generation

SynFAGnet: A Fully Automated Generative Network for Realistic Fire Image Generation

This paper proposes a fully automated generative network (“SynFAGnet”) for automatically creating a realistic-looking synthetic fire image. SynFAGnet is used as a data augmentation technique to create diverse data for training models, thereby solving problems related to real data acquisition and data imbalances. SynFAGnet comprises two main parts: an object-scene placement net (OSPNet) and a local–global context-based generative adversarial network (LGC-GAN). The OSPNet identifies suitable positions and scales for fires corresponding to the background scene. The LGC-GAN enhances the realistic appearance of synthetic fire images created by a given fire object-background scene pair by assembling effects such as halos and reflections in the surrounding area in the background scene. A comparative analysis shows that SynFAGnet achieves better outcomes than previous studies for both the Fréchet inception distance and learned perceptual image patch similarity evaluation metrics (values of 17.232 and 0.077, respectively). In addition, SynFAGnet is verified as a practically applicable data augmentation technique for training datasets, as it improves the detection and instance segmentation performance.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Fire Technology
Fire Technology 工程技术-材料科学:综合
CiteScore
6.60
自引率
14.70%
发文量
137
审稿时长
7.5 months
期刊介绍: Fire Technology publishes original contributions, both theoretical and empirical, that contribute to the solution of problems in fire safety science and engineering. It is the leading journal in the field, publishing applied research dealing with the full range of actual and potential fire hazards facing humans and the environment. It covers the entire domain of fire safety science and engineering problems relevant in industrial, operational, cultural, and environmental applications, including modeling, testing, detection, suppression, human behavior, wildfires, structures, and risk analysis. The aim of Fire Technology is to push forward the frontiers of knowledge and technology by encouraging interdisciplinary communication of significant technical developments in fire protection and subjects of scientific interest to the fire protection community at large. It is published in conjunction with the National Fire Protection Association (NFPA) and the Society of Fire Protection Engineers (SFPE). The mission of NFPA is to help save lives and reduce loss with information, knowledge, and passion. The mission of SFPE is advancing the science and practice of fire protection engineering internationally.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信