Special issue on autonomous unmanned aerial/ground vehicles and their applications

IF 1.3 4区 计算机科学 Q3 ENGINEERING, ELECTRICAL & ELECTRONIC
ETRI Journal Pub Date : 2023-10-29 DOI:10.4218/etr2.12628
Joongheon Kim, Yu-Cheol Lee, Jun Hwan Lee, Jin Seek Choi
{"title":"Special issue on autonomous unmanned aerial/ground vehicles and their applications","authors":"Joongheon Kim,&nbsp;Yu-Cheol Lee,&nbsp;Jun Hwan Lee,&nbsp;Jin Seek Choi","doi":"10.4218/etr2.12628","DOIUrl":null,"url":null,"abstract":"<p>Recently, research on autonomous mobility control has been actively and widely conducted for various applications. In particular, autonomous mobility control for unmanned aerial and ground vehicles has been our research interest because it has considerable challenges, such as time-consuming and high-delay computations, complicated functionalities, and dangerous tasks that were previously performed by humans. Furthermore, fully autonomous unmanned aerial and ground vehicles are barely practical and have various operational limitations, such as high-precision sensing, high computational complexity, low autonomy, and restricted mobility. To develop the required technologies to overcome these limitations and achieve full autonomy for unmanned aerial and ground vehicles, various studies have addressed aspects such as precise pose estimation, environment mapping, path planning, trajectory optimization, and 2D/3D object tracking and detection.</p><p>With fully autonomous operation and functionalities for unmanned aerial and ground vehicles, emerging applications will become more diverse and include autonomous artificial-intelligence-based surveillance, autonomous disaster prevention broadcasting and control, mobile autonomous aerial and ground wireless/cellular access service provisioning, autonomous multirobot coordination, and cooperation for smart factory management in smart city applications, for which a skilled human operator must currently intervene throughout operation.</p><p>For this special issue, we selected 11 key studies on (1) communication, networks, and mobility [<span>1-5</span>] and (2) object detection and tracking in autonomous driving [<span>6-11</span>].</p><p>In [<span>1</span>], surveys and discussions are presented on recent deep-learning-based developments to achieve autonomous mobility control and efficient resource management of autonomous vehicles including unmanned aerial vehicles (UAVs). The developments include multiagent reinforcement learning and neural Myerson auction. We believe that integrating multiagent reinforcement learning and neural Myerson auction will be critical for efficient and trustworthy autonomous mobility services.</p><p>In [<span>2</span>], a safe landing algorithm is introduced for urban drone delivery. The proposed algorithm generates a safe and efficient vertical landing path for drones, allowing them to avoid obstacles commonly found in urban environments, such as trees, streetlights, utility poles, and wires. To this end, landing-angle control is implemented to land vertically, and a rapidly-exploring random tree (RRT) is used in a collision avoidance algorithm. This combination of methods enables precise and reliable drone delivery in urban settings.</p><p>In [<span>3</span>], a loosely coupled relative position estimation method is proposed based on a decentralized ultrawideband global navigation support system and inertial navigation system for flight controllers. Key obstacles to multi-drone collaboration are noted and include relative positional errors and the absence of communication devices. To address such problems, an extended Kalman filter (EKF) is adopted to correct distance errors by fusing ultrawideband data acquired through random communications using a novel UWB communication module.</p><p>Unmanned vehicles are being increasingly used for time-consuming, complicated, and dangerous tasks that were previously performed by humans. However, they have limitations for applications like establishing high-speed wireless networks. In [<span>4</span>], a 3D geometry-based stochastic model for UAV multiple-input multiple-output (MIMO) channels is presented. The UAV flying direction and location have a significant impact on MIMO performance. This innovative model of 3D navigation and scattering environments is closely related to the scope of this special issue.</p><p>In [<span>5</span>], UAVs are considered essential components in non-terrestrial networks, especially in 5G-and-beyond communication systems. Employing UAVs operated in conjunction with a 4G/5G base station has proven to be a practical solution for providing cellular network services in areas where conventional communication infrastructures are unavailable. This paper introduces the uncrewed aerial vehicle–base station system that utilizes a high-capacity wireless backhaul operating in millimeter wave frequency bands.</p><p>In [<span>6</span>], advanced video analytics for tasks such as moving object detection and segmentation are presented, thereby increasing the demand for such methods in unmanned aerial and ground vehicle applications. A novel zero-shot video object segmentation is introduced to focus on the discovery of moving objects in challenging scenarios. This method employs a background memory model for training from sparse annotations over time by using temporal modeling of the background to accurately detect moving objects. In addition, the method addresses the limitations of existing state-of-the-art solutions for detecting salient objects within images regardless of their motion.</p><p>In [<span>7</span>], an adaptive UAV-assisted object-recognition algorithm is introduced for urban surveillance scenarios. In a UAV-assisted surveillance system, UAVs are equipped with learning-based object recognition models and can collect surveillance images. Owing to UAV limitations (for example, limited battery and computational capabilities), adaptive control considering these limitations is devised to maximize the time-averaged recognition performance subject to stability through Lyapunov optimization.</p><p>In [<span>8</span>], modern semantic segmentation frameworks combining low- and high-level context information are used to improve performance. In addition, post-level context information is considered in a context refinement network (CRFNet). Training for improving the semantic segmentation predictions proceeds through an encoder–decoder structure. This study directly considers the relation between spatially neighboring pixels of a label map using methods such as Markov and conditional random fields.</p><p>In [<span>9</span>], real-time accurate 3D multi-pedestrian detection and tracking are achieved using 3D LiDAR point clouds from crowded environments. Pedestrian detection segments a sparse 3D point cloud into individual pedestrians using a lightweight convolutional autoencoder and connected component labeling. Multi-pedestrian tracking associates the same pedestrians by considering motion and appearance cues in continuous frames. In addition, the dynamic movements of pedestrians are estimated with various patterns by adaptively mixing heterogeneous motion models.</p><p>In [<span>10</span>], sensor-fusion-based object detection and classification are presented. The proposed method operates in real time, rendering it suitable for integration into autonomous vehicles. It performs well on a custom dataset and publicly available datasets, demonstrating its effectiveness in real-world road environments. In addition, a 3D moving object detection dataset called ETRI 3D MOD, is constructed.</p><p>In [<span>11</span>], three techniques for combining information from multiple cameras are proposed, namely, feature, early, and late fusion techniques. Extensive experiments were conducted on pedestrian-view intersection classification. The proposed model with feature fusion provides an area under the curve and an F1-score of 82.00 and 46.48, respectively, outperforming a model trained using only real three-camera data and one-camera models by a large margin.</p><p>The authors declare that there are no conflicts of interest.</p>","PeriodicalId":11901,"journal":{"name":"ETRI Journal","volume":"45 5","pages":"731-734"},"PeriodicalIF":1.3000,"publicationDate":"2023-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ETRI Journal","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.4218/etr2.12628","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

Recently, research on autonomous mobility control has been actively and widely conducted for various applications. In particular, autonomous mobility control for unmanned aerial and ground vehicles has been our research interest because it has considerable challenges, such as time-consuming and high-delay computations, complicated functionalities, and dangerous tasks that were previously performed by humans. Furthermore, fully autonomous unmanned aerial and ground vehicles are barely practical and have various operational limitations, such as high-precision sensing, high computational complexity, low autonomy, and restricted mobility. To develop the required technologies to overcome these limitations and achieve full autonomy for unmanned aerial and ground vehicles, various studies have addressed aspects such as precise pose estimation, environment mapping, path planning, trajectory optimization, and 2D/3D object tracking and detection.

With fully autonomous operation and functionalities for unmanned aerial and ground vehicles, emerging applications will become more diverse and include autonomous artificial-intelligence-based surveillance, autonomous disaster prevention broadcasting and control, mobile autonomous aerial and ground wireless/cellular access service provisioning, autonomous multirobot coordination, and cooperation for smart factory management in smart city applications, for which a skilled human operator must currently intervene throughout operation.

For this special issue, we selected 11 key studies on (1) communication, networks, and mobility [1-5] and (2) object detection and tracking in autonomous driving [6-11].

In [1], surveys and discussions are presented on recent deep-learning-based developments to achieve autonomous mobility control and efficient resource management of autonomous vehicles including unmanned aerial vehicles (UAVs). The developments include multiagent reinforcement learning and neural Myerson auction. We believe that integrating multiagent reinforcement learning and neural Myerson auction will be critical for efficient and trustworthy autonomous mobility services.

In [2], a safe landing algorithm is introduced for urban drone delivery. The proposed algorithm generates a safe and efficient vertical landing path for drones, allowing them to avoid obstacles commonly found in urban environments, such as trees, streetlights, utility poles, and wires. To this end, landing-angle control is implemented to land vertically, and a rapidly-exploring random tree (RRT) is used in a collision avoidance algorithm. This combination of methods enables precise and reliable drone delivery in urban settings.

In [3], a loosely coupled relative position estimation method is proposed based on a decentralized ultrawideband global navigation support system and inertial navigation system for flight controllers. Key obstacles to multi-drone collaboration are noted and include relative positional errors and the absence of communication devices. To address such problems, an extended Kalman filter (EKF) is adopted to correct distance errors by fusing ultrawideband data acquired through random communications using a novel UWB communication module.

Unmanned vehicles are being increasingly used for time-consuming, complicated, and dangerous tasks that were previously performed by humans. However, they have limitations for applications like establishing high-speed wireless networks. In [4], a 3D geometry-based stochastic model for UAV multiple-input multiple-output (MIMO) channels is presented. The UAV flying direction and location have a significant impact on MIMO performance. This innovative model of 3D navigation and scattering environments is closely related to the scope of this special issue.

In [5], UAVs are considered essential components in non-terrestrial networks, especially in 5G-and-beyond communication systems. Employing UAVs operated in conjunction with a 4G/5G base station has proven to be a practical solution for providing cellular network services in areas where conventional communication infrastructures are unavailable. This paper introduces the uncrewed aerial vehicle–base station system that utilizes a high-capacity wireless backhaul operating in millimeter wave frequency bands.

In [6], advanced video analytics for tasks such as moving object detection and segmentation are presented, thereby increasing the demand for such methods in unmanned aerial and ground vehicle applications. A novel zero-shot video object segmentation is introduced to focus on the discovery of moving objects in challenging scenarios. This method employs a background memory model for training from sparse annotations over time by using temporal modeling of the background to accurately detect moving objects. In addition, the method addresses the limitations of existing state-of-the-art solutions for detecting salient objects within images regardless of their motion.

In [7], an adaptive UAV-assisted object-recognition algorithm is introduced for urban surveillance scenarios. In a UAV-assisted surveillance system, UAVs are equipped with learning-based object recognition models and can collect surveillance images. Owing to UAV limitations (for example, limited battery and computational capabilities), adaptive control considering these limitations is devised to maximize the time-averaged recognition performance subject to stability through Lyapunov optimization.

In [8], modern semantic segmentation frameworks combining low- and high-level context information are used to improve performance. In addition, post-level context information is considered in a context refinement network (CRFNet). Training for improving the semantic segmentation predictions proceeds through an encoder–decoder structure. This study directly considers the relation between spatially neighboring pixels of a label map using methods such as Markov and conditional random fields.

In [9], real-time accurate 3D multi-pedestrian detection and tracking are achieved using 3D LiDAR point clouds from crowded environments. Pedestrian detection segments a sparse 3D point cloud into individual pedestrians using a lightweight convolutional autoencoder and connected component labeling. Multi-pedestrian tracking associates the same pedestrians by considering motion and appearance cues in continuous frames. In addition, the dynamic movements of pedestrians are estimated with various patterns by adaptively mixing heterogeneous motion models.

In [10], sensor-fusion-based object detection and classification are presented. The proposed method operates in real time, rendering it suitable for integration into autonomous vehicles. It performs well on a custom dataset and publicly available datasets, demonstrating its effectiveness in real-world road environments. In addition, a 3D moving object detection dataset called ETRI 3D MOD, is constructed.

In [11], three techniques for combining information from multiple cameras are proposed, namely, feature, early, and late fusion techniques. Extensive experiments were conducted on pedestrian-view intersection classification. The proposed model with feature fusion provides an area under the curve and an F1-score of 82.00 and 46.48, respectively, outperforming a model trained using only real three-camera data and one-camera models by a large margin.

The authors declare that there are no conflicts of interest.

关于无人驾驶飞行器及其应用的特刊
在[7]中,针对城市监控场景,引入了一种自适应无人机辅助目标识别算法。在无人机辅助监视系统中,无人机配备了基于学习的目标识别模型,可以收集监视图像。由于无人机的局限性(例如,有限的电池和计算能力),考虑到这些局限性,设计了自适应控制,以通过李雅普诺夫优化最大限度地提高稳定性下的时间平均识别性能。在[8]中,结合低级和高级上下文信息的现代语义分割框架被用于提高性能。此外,在上下文细化网络(CRFNet)中考虑了后级上下文信息。用于改进语义分割预测的训练通过编码器-解码器结构进行。本研究使用马尔可夫和条件随机场等方法直接考虑标签图的空间相邻像素之间的关系。在[9]中,使用拥挤环境中的3D激光雷达点云实现了实时精确的3D多行人检测和跟踪。行人检测使用轻量级卷积自动编码器和连接组件标记将稀疏的3D点云分割成单个行人。多行人跟踪通过考虑连续帧中的运动和外观线索来关联相同的行人。此外,通过自适应地混合异构运动模型,以各种模式估计行人的动态运动。在[10]中,提出了基于传感器融合的目标检测和分类。所提出的方法实时运行,使其适合集成到自动驾驶汽车中。它在自定义数据集和公开数据集上表现良好,证明了它在现实道路环境中的有效性。此外,还构建了一个名为ETRI3DMOD的三维运动物体检测数据集。在[11]中,提出了三种用于组合来自多个相机的信息的技术,即特征、早期和晚期融合技术。对行人视野交叉口的分类进行了广泛的实验。所提出的具有特征融合的模型分别提供了82.00和46.48的曲线下面积和F1分数,大大优于仅使用真实三相机数据和一相机模型训练的模型。提交人声明不存在利益冲突。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
ETRI Journal
ETRI Journal 工程技术-电信学
CiteScore
4.00
自引率
7.10%
发文量
98
审稿时长
6.9 months
期刊介绍: ETRI Journal is an international, peer-reviewed multidisciplinary journal published bimonthly in English. The main focus of the journal is to provide an open forum to exchange innovative ideas and technology in the fields of information, telecommunications, and electronics. Key topics of interest include high-performance computing, big data analytics, cloud computing, multimedia technology, communication networks and services, wireless communications and mobile computing, material and component technology, as well as security. With an international editorial committee and experts from around the world as reviewers, ETRI Journal publishes high-quality research papers on the latest and best developments from the global community.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信