Le Hoang Duong, T. T. Huynh, Minh Tam Pham, Gwangzeen Ko, Jung Ick Moon, Jun Jo, N. Q. Hung
{"title":"面向自动驾驶机器人的轻量级目标检测框架","authors":"Le Hoang Duong, T. T. Huynh, Minh Tam Pham, Gwangzeen Ko, Jung Ick Moon, Jun Jo, N. Q. Hung","doi":"10.1109/DICTA52665.2021.9647256","DOIUrl":null,"url":null,"abstract":"Object detection is an emerging and essential problem in recent years, which has been widely applied in many aspects of daily life such as video surveillance, self-driving robots, and automatic payment. The rapid development of deep learning models allows object detectors to work in real-time with high accuracy. However, such a sophisticated model often requires robust computing infrastructure such as powerful graphics processing units (GPUs). This requirement might cause a severe issue for embedded systems with small, power-efficient artificial intelligence (AI) systems like Jetson Nano, which are often restricted in both memory storage and computing sheer power. In this work, we aim to address this challenge by proposing a lightweight object detection framework that is specialized for the Internet of Things (IoT) devices with low-power processors such as Jetson Nano. In order to detect the object with different size, our framework employs a backbone residual CNN-based network as the feature extractor. We then design a multi-layer model to combine the feature at different levels of granularity, before using the processed feature to locate and classify the object. We also apply augmentation techniques to enhance the robustness of the framework to adversarial factors. Extensive experiments on real devices in many scenarios, such as autonomous cars or wireless robot recharging systems, showed that our technique can achieve nearly on par results with the state-of-the-art YOLOv5 while requires only one-fourth of computation power.","PeriodicalId":424950,"journal":{"name":"2021 Digital Image Computing: Techniques and Applications (DICTA)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"ODAR: A Lightweight Object Detection Framework for Autonomous Driving Robots\",\"authors\":\"Le Hoang Duong, T. T. Huynh, Minh Tam Pham, Gwangzeen Ko, Jung Ick Moon, Jun Jo, N. Q. Hung\",\"doi\":\"10.1109/DICTA52665.2021.9647256\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Object detection is an emerging and essential problem in recent years, which has been widely applied in many aspects of daily life such as video surveillance, self-driving robots, and automatic payment. The rapid development of deep learning models allows object detectors to work in real-time with high accuracy. However, such a sophisticated model often requires robust computing infrastructure such as powerful graphics processing units (GPUs). This requirement might cause a severe issue for embedded systems with small, power-efficient artificial intelligence (AI) systems like Jetson Nano, which are often restricted in both memory storage and computing sheer power. In this work, we aim to address this challenge by proposing a lightweight object detection framework that is specialized for the Internet of Things (IoT) devices with low-power processors such as Jetson Nano. In order to detect the object with different size, our framework employs a backbone residual CNN-based network as the feature extractor. We then design a multi-layer model to combine the feature at different levels of granularity, before using the processed feature to locate and classify the object. We also apply augmentation techniques to enhance the robustness of the framework to adversarial factors. Extensive experiments on real devices in many scenarios, such as autonomous cars or wireless robot recharging systems, showed that our technique can achieve nearly on par results with the state-of-the-art YOLOv5 while requires only one-fourth of computation power.\",\"PeriodicalId\":424950,\"journal\":{\"name\":\"2021 Digital Image Computing: Techniques and Applications (DICTA)\",\"volume\":\"26 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 Digital Image Computing: Techniques and Applications (DICTA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/DICTA52665.2021.9647256\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 Digital Image Computing: Techniques and Applications (DICTA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DICTA52665.2021.9647256","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
ODAR: A Lightweight Object Detection Framework for Autonomous Driving Robots
Object detection is an emerging and essential problem in recent years, which has been widely applied in many aspects of daily life such as video surveillance, self-driving robots, and automatic payment. The rapid development of deep learning models allows object detectors to work in real-time with high accuracy. However, such a sophisticated model often requires robust computing infrastructure such as powerful graphics processing units (GPUs). This requirement might cause a severe issue for embedded systems with small, power-efficient artificial intelligence (AI) systems like Jetson Nano, which are often restricted in both memory storage and computing sheer power. In this work, we aim to address this challenge by proposing a lightweight object detection framework that is specialized for the Internet of Things (IoT) devices with low-power processors such as Jetson Nano. In order to detect the object with different size, our framework employs a backbone residual CNN-based network as the feature extractor. We then design a multi-layer model to combine the feature at different levels of granularity, before using the processed feature to locate and classify the object. We also apply augmentation techniques to enhance the robustness of the framework to adversarial factors. Extensive experiments on real devices in many scenarios, such as autonomous cars or wireless robot recharging systems, showed that our technique can achieve nearly on par results with the state-of-the-art YOLOv5 while requires only one-fourth of computation power.