An end to end workflow for synthetic data generation for robust object detection*

Johannes Metzler, Fouad Bahrpeyma, Dirk Reichelt
{"title":"An end to end workflow for synthetic data generation for robust object detection*","authors":"Johannes Metzler, Fouad Bahrpeyma, Dirk Reichelt","doi":"10.1109/INDIN51400.2023.10218035","DOIUrl":null,"url":null,"abstract":"Object detection is a task in computer vision that involves detecting instances of visual objects of a particular class in digital images. Numerous computer vision tasks highly depend on object detection such as instance segmentation, image captioning and object tracking. A major purpose of object detection is to develop computational models that provide inputs crucial to computer vision applications. Convolutional Neural Networks (CNNs) have recently become popular due to their key roles in enabling object detection. However, the performance of CNNs is largely dependent upon the quality and quantity of training datasets, which are often difficult to obtain in real-world applications. In order to ensure the robustness of such models, it is vital that training instances are provided under various randomized conditions. These conditions are typically a combination of a variety of factors, including lighting conditions, object location, the presence of multiple objects in the scene, varieties of backgrounds, and the angle of the camera. In particular, companies, depending on their applications (such as fault detection, anomaly detection, condition monitoring, predictive quality and so on), require specialized models for their custom products and so always face difficulties in providing a large number of randomized conditioned instances of their objects. The primary reason is that the process of capturing randomized conditioned images of real objects is usually costly, time-consuming, and challenging in practice. Due to the efficiency gained so far via the use of synthetic data for training such systems, synthetic data has recently attracted considerable attention. This paper presents an end-to-end synthetic data generation method for building a robust object detection model for customized products using NVIDIA Omniverse and CNNs. In this paper, we demonstrate and evaluate our contribution to the modeling of chess pieces, where a total accuracy of 98.8 % was obtained.","PeriodicalId":174443,"journal":{"name":"2023 IEEE 21st International Conference on Industrial Informatics (INDIN)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE 21st International Conference on Industrial Informatics (INDIN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/INDIN51400.2023.10218035","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Object detection is a task in computer vision that involves detecting instances of visual objects of a particular class in digital images. Numerous computer vision tasks highly depend on object detection such as instance segmentation, image captioning and object tracking. A major purpose of object detection is to develop computational models that provide inputs crucial to computer vision applications. Convolutional Neural Networks (CNNs) have recently become popular due to their key roles in enabling object detection. However, the performance of CNNs is largely dependent upon the quality and quantity of training datasets, which are often difficult to obtain in real-world applications. In order to ensure the robustness of such models, it is vital that training instances are provided under various randomized conditions. These conditions are typically a combination of a variety of factors, including lighting conditions, object location, the presence of multiple objects in the scene, varieties of backgrounds, and the angle of the camera. In particular, companies, depending on their applications (such as fault detection, anomaly detection, condition monitoring, predictive quality and so on), require specialized models for their custom products and so always face difficulties in providing a large number of randomized conditioned instances of their objects. The primary reason is that the process of capturing randomized conditioned images of real objects is usually costly, time-consuming, and challenging in practice. Due to the efficiency gained so far via the use of synthetic data for training such systems, synthetic data has recently attracted considerable attention. This paper presents an end-to-end synthetic data generation method for building a robust object detection model for customized products using NVIDIA Omniverse and CNNs. In this paper, we demonstrate and evaluate our contribution to the modeling of chess pieces, where a total accuracy of 98.8 % was obtained.
一个端到端工作流合成数据生成鲁棒对象检测*
对象检测是计算机视觉中的一项任务,涉及检测数字图像中特定类别的视觉对象的实例。许多计算机视觉任务高度依赖于目标检测,如实例分割、图像字幕和目标跟踪。目标检测的一个主要目的是开发为计算机视觉应用提供关键输入的计算模型。卷积神经网络(cnn)最近因其在实现目标检测方面的关键作用而受到欢迎。然而,cnn的性能很大程度上取决于训练数据集的质量和数量,而这些数据集在实际应用中往往很难获得。为了保证模型的鲁棒性,在各种随机条件下提供训练实例是至关重要的。这些条件通常是各种因素的组合,包括照明条件、物体位置、场景中多个物体的存在、各种背景和相机的角度。特别是,公司,根据他们的应用程序(如故障检测、异常检测、状态监控、预测质量等),需要为他们的定制产品提供专门的模型,因此在提供大量随机条件的对象实例时总是面临困难。主要原因是捕获真实对象的随机条件图像的过程通常是昂贵的,耗时的,并且在实践中具有挑战性。由于迄今为止通过使用合成数据来训练此类系统所获得的效率,合成数据最近引起了相当大的关注。本文提出了一种基于NVIDIA Omniverse和cnn的端到端合成数据生成方法,用于构建定制产品的鲁棒目标检测模型。在本文中,我们展示并评估了我们对棋子建模的贡献,其中获得了98.8%的总准确率。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信