Nilesh Aggarwal, Anunay, Vayam Jain, Tushar Singh, D. Vishwakarma
{"title":"DLVS: Time Series Architecture for Image-Based Visual Servoing","authors":"Nilesh Aggarwal, Anunay, Vayam Jain, Tushar Singh, D. Vishwakarma","doi":"10.1109/ICCRE57112.2023.10155585","DOIUrl":null,"url":null,"abstract":"A novel deep learning-based visual servoing architecture “DLVS” is proposed for control of an unmanned aerial vehicle (UAV) capable of quasi-stationary flight with a camera mounted under the vehicle to track a target consisting of a finite set of stationary points lying in a plane. Current Deep Learning and Reinforcement Learning (RL) based end-to-end servoing approaches rely on training convolutional neural networks using color images with known camera poses to learn the visual features in the environment suitable for servoing tasks. This approach limits the application of the network to available environments where the dataset was collected. Moreover, we cannot deploy such networks on the low-power computers present onboard the UAV. The proposed solution employs a time series architecture to learn temporal data from sequential values to output the control cues to the flight controller. The low computational complexity and flexibility of the DLVS architecture ensure real-time onboard tracking for virtually any target. The algorithm was thoroughly validated in real-life environments and outperformed the current state-of-the-art in terms of time efficiency and accuracy.","PeriodicalId":285164,"journal":{"name":"2023 8th International Conference on Control and Robotics Engineering (ICCRE)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 8th International Conference on Control and Robotics Engineering (ICCRE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCRE57112.2023.10155585","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
A novel deep learning-based visual servoing architecture “DLVS” is proposed for control of an unmanned aerial vehicle (UAV) capable of quasi-stationary flight with a camera mounted under the vehicle to track a target consisting of a finite set of stationary points lying in a plane. Current Deep Learning and Reinforcement Learning (RL) based end-to-end servoing approaches rely on training convolutional neural networks using color images with known camera poses to learn the visual features in the environment suitable for servoing tasks. This approach limits the application of the network to available environments where the dataset was collected. Moreover, we cannot deploy such networks on the low-power computers present onboard the UAV. The proposed solution employs a time series architecture to learn temporal data from sequential values to output the control cues to the flight controller. The low computational complexity and flexibility of the DLVS architecture ensure real-time onboard tracking for virtually any target. The algorithm was thoroughly validated in real-life environments and outperformed the current state-of-the-art in terms of time efficiency and accuracy.