{"title":"基于虚拟环境的神经网络训练图像生成工具的开发","authors":"R. Arenas, P. Delmas, Alfonso Gastelum-Strozzi","doi":"10.1109/IVCNZ51579.2020.9290491","DOIUrl":null,"url":null,"abstract":"We present a computational tool to generate visual and descriptive data used as additional training images for neural networks involved in image recognition tasks. The work is inspired by the problem posed to acquire enough data, in order to train service robots, with the goal of improving the range of objects in the environment with which they can interact. The tool provides a framework that allows users to easily setup different environments with the visual information needed for the training, accordingly to their needs. The tool was developed with the Unity engine, and it was designed to be able to import external prefabs. These models are standardized and catalogued into lists, which are accessed to create more complex and diverse virtual environments. Another component of the tool adds an additional layer of complexity by creating randomized environments with different conditions (scale, position and orientation of objects, and environmental illumination). The performance of the created dataset was tested by training the information on the YOLO-V3 (You Only Look Once) architecture and testing on both artificial and real images.","PeriodicalId":164317,"journal":{"name":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","volume":"118 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Development of a Virtual Environment Based Image Generation Tool for Neural Network Training\",\"authors\":\"R. Arenas, P. Delmas, Alfonso Gastelum-Strozzi\",\"doi\":\"10.1109/IVCNZ51579.2020.9290491\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We present a computational tool to generate visual and descriptive data used as additional training images for neural networks involved in image recognition tasks. The work is inspired by the problem posed to acquire enough data, in order to train service robots, with the goal of improving the range of objects in the environment with which they can interact. The tool provides a framework that allows users to easily setup different environments with the visual information needed for the training, accordingly to their needs. The tool was developed with the Unity engine, and it was designed to be able to import external prefabs. These models are standardized and catalogued into lists, which are accessed to create more complex and diverse virtual environments. Another component of the tool adds an additional layer of complexity by creating randomized environments with different conditions (scale, position and orientation of objects, and environmental illumination). The performance of the created dataset was tested by training the information on the YOLO-V3 (You Only Look Once) architecture and testing on both artificial and real images.\",\"PeriodicalId\":164317,\"journal\":{\"name\":\"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)\",\"volume\":\"118 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-11-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IVCNZ51579.2020.9290491\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 35th International Conference on Image and Vision Computing New Zealand (IVCNZ)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IVCNZ51579.2020.9290491","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
我们提出了一种计算工具,用于生成视觉和描述性数据,作为涉及图像识别任务的神经网络的额外训练图像。这项工作的灵感来自于一个问题,即获取足够的数据,以训练服务机器人,目标是提高环境中物体的范围,使它们能够与之互动。该工具提供了一个框架,允许用户根据自己的需要轻松地设置不同的环境,并提供培训所需的视觉信息。该工具是使用Unity引擎开发的,它被设计成能够导入外部预制件。这些模型被标准化并编目到列表中,可以访问这些列表来创建更复杂和多样化的虚拟环境。该工具的另一个组件通过创建具有不同条件(对象的规模、位置和方向以及环境照明)的随机环境增加了额外的复杂性层。通过在YOLO-V3 (You Only Look Once)架构上训练信息,并在人工图像和真实图像上进行测试,测试了所创建数据集的性能。
Development of a Virtual Environment Based Image Generation Tool for Neural Network Training
We present a computational tool to generate visual and descriptive data used as additional training images for neural networks involved in image recognition tasks. The work is inspired by the problem posed to acquire enough data, in order to train service robots, with the goal of improving the range of objects in the environment with which they can interact. The tool provides a framework that allows users to easily setup different environments with the visual information needed for the training, accordingly to their needs. The tool was developed with the Unity engine, and it was designed to be able to import external prefabs. These models are standardized and catalogued into lists, which are accessed to create more complex and diverse virtual environments. Another component of the tool adds an additional layer of complexity by creating randomized environments with different conditions (scale, position and orientation of objects, and environmental illumination). The performance of the created dataset was tested by training the information on the YOLO-V3 (You Only Look Once) architecture and testing on both artificial and real images.