基于视觉语义分割和辅助分类器的资源约束边缘系统场景识别

Chungjae Choe, Sungwook Jung, Nak-Myoung Sung, Sukjun Lee
{"title":"基于视觉语义分割和辅助分类器的资源约束边缘系统场景识别","authors":"Chungjae Choe, Sungwook Jung, Nak-Myoung Sung, Sukjun Lee","doi":"10.1109/icoin56518.2023.10048947","DOIUrl":null,"url":null,"abstract":"This paper presents a scene identification method employing semantic segmentation where the method provides real-time computation in resource-constrained edge devices. Scene identification could be crucial for intelligent systems (e.g., service robots, drone-based inspection, and visual surveillance) regarding a proper decision making of those systems. Existing methods focus on adopting a deep learning-based image classification for the identification. However, those approaches may provide wrong identification due to an overlap of spatial features when training dataset is limited.In this paper, we propose an accurate scene identification with a novel approach. Our method includes two-steps: 1) measurement of object class frequency with visual semantic segmentation; 2) scene classification using class frequencies. For fast computation, we build a lightweight backbone network for the segmentation model in addition to TensorRT-based optimization. From the experiments, we validate that our method improves the identification accuracy by 12% compared to conventional visual classification-based method. In terms of computation, we observe that the method enables real-time inference on resource- constrained devices (i.e., NVIDIA Jetsons).","PeriodicalId":285763,"journal":{"name":"2023 International Conference on Information Networking (ICOIN)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Scene identification using visual semantic segmentation and supplementary classifier for resource-constrained edge systems\",\"authors\":\"Chungjae Choe, Sungwook Jung, Nak-Myoung Sung, Sukjun Lee\",\"doi\":\"10.1109/icoin56518.2023.10048947\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper presents a scene identification method employing semantic segmentation where the method provides real-time computation in resource-constrained edge devices. Scene identification could be crucial for intelligent systems (e.g., service robots, drone-based inspection, and visual surveillance) regarding a proper decision making of those systems. Existing methods focus on adopting a deep learning-based image classification for the identification. However, those approaches may provide wrong identification due to an overlap of spatial features when training dataset is limited.In this paper, we propose an accurate scene identification with a novel approach. Our method includes two-steps: 1) measurement of object class frequency with visual semantic segmentation; 2) scene classification using class frequencies. For fast computation, we build a lightweight backbone network for the segmentation model in addition to TensorRT-based optimization. From the experiments, we validate that our method improves the identification accuracy by 12% compared to conventional visual classification-based method. In terms of computation, we observe that the method enables real-time inference on resource- constrained devices (i.e., NVIDIA Jetsons).\",\"PeriodicalId\":285763,\"journal\":{\"name\":\"2023 International Conference on Information Networking (ICOIN)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 International Conference on Information Networking (ICOIN)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/icoin56518.2023.10048947\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference on Information Networking (ICOIN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/icoin56518.2023.10048947","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

本文提出了一种基于语义分割的场景识别方法,该方法在资源受限的边缘设备上提供了实时计算。场景识别对于智能系统(例如,服务机器人、基于无人机的检查和视觉监控)来说,对于这些系统的正确决策至关重要。现有的方法主要是采用基于深度学习的图像分类进行识别。然而,当训练数据集有限时,这些方法可能会由于空间特征的重叠而提供错误的识别。在本文中,我们提出了一种新的精确场景识别方法。该方法分为两步:1)基于视觉语义分割的目标类频率测量;2)使用类频率进行场景分类。为了提高分割模型的计算速度,除了基于tensorrt的优化外,我们还为分割模型构建了一个轻量级的骨干网络。实验结果表明,与传统的基于视觉分类的方法相比,该方法的识别准确率提高了12%。在计算方面,我们观察到该方法可以在资源受限的设备(即NVIDIA Jetsons)上进行实时推理。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Scene identification using visual semantic segmentation and supplementary classifier for resource-constrained edge systems
This paper presents a scene identification method employing semantic segmentation where the method provides real-time computation in resource-constrained edge devices. Scene identification could be crucial for intelligent systems (e.g., service robots, drone-based inspection, and visual surveillance) regarding a proper decision making of those systems. Existing methods focus on adopting a deep learning-based image classification for the identification. However, those approaches may provide wrong identification due to an overlap of spatial features when training dataset is limited.In this paper, we propose an accurate scene identification with a novel approach. Our method includes two-steps: 1) measurement of object class frequency with visual semantic segmentation; 2) scene classification using class frequencies. For fast computation, we build a lightweight backbone network for the segmentation model in addition to TensorRT-based optimization. From the experiments, we validate that our method improves the identification accuracy by 12% compared to conventional visual classification-based method. In terms of computation, we observe that the method enables real-time inference on resource- constrained devices (i.e., NVIDIA Jetsons).
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信