Human Detection and Action Recognition for Search and Rescue in Disasters Using YOLOv3 Algorithm

B. Valarmathi, Jain Kshitij, Rajpurohit Dimple, N. Gupta, Y. H. Robinson, G. Arulkumaran, Tadesse Mulu
{"title":"Human Detection and Action Recognition for Search and Rescue in Disasters Using YOLOv3 Algorithm","authors":"B. Valarmathi, Jain Kshitij, Rajpurohit Dimple, N. Gupta, Y. H. Robinson, G. Arulkumaran, Tadesse Mulu","doi":"10.1155/2023/5419384","DOIUrl":null,"url":null,"abstract":"Drone examination has been overall quickly embraced by NDMM (natural disaster mitigation and management) division to survey the state of impacted regions. Manual video analysis by human observers takes time and is subject to mistakes. The human identification examination of pictures caught by drones will give a practical method for saving lives who are being trapped under debris during quakes or in floods and so on. Drone investigation for research and security and search and rescue (SAR) should involve the drone to filter the impacted area using a camera and a model of unmanned area vehicles (UAVs) to identify specific locations where assistance is required. The existing methods (Balmukund et al. 2020) used were faster-region based convolutional neural networks (F-RCNNs), single shot detector (SSD), and region-based fully convolutional network (R-FCN) for the detection of human and recognition of action. Some of the existing methods used 700 images with six classes only, whereas the proposed model uses 1996 images with eight classes. The proposed model is used YOLOv3 (you only look once) algorithm for the detection and recognition of actions. In this study, we provide the fundamental ideas underlying an object detection model. To find the most effective model for human recognition and detection, we trained the YOLOv3 algorithm on the image dataset and evaluated its performance. We compared the outcomes with the existing algorithms like F-RCNN, SSD, and R-FCN. The accuracies of F-RCNN, SSD, R-FCN (existing algorithms), and YOLOv3 (proposed algorithm) are 53%, 73%, 93%, and 94.9%, respectively. Among these algorithms, the YOLOv3 algorithm gives the highest accuracy of 94.9%. The proposed work shows that existing models are inadequate for critical applications like search and rescue, which convinces us to propose a model raised by a pyramidal component extracting SSD in human localization and action recognition. The suggested model is 94.9% accurate when applied to the proposed dataset, which is an important contribution. Likewise, the suggested model succeeds in helping time for expectation in examination with the cutting-edge identification models with existing strategies. The average time taken by our proposed technique to distinguish a picture is 0.40 milisec which is a lot better than the existing method. The proposed model can likewise distinguish video and can be utilized for real-time recognition. The SSD model can likewise use to anticipate messages if present in the picture.","PeriodicalId":23352,"journal":{"name":"Turkish J. Electr. Eng. Comput. Sci.","volume":"2 1","pages":"5419384:1-5419384:19"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Turkish J. Electr. Eng. Comput. Sci.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1155/2023/5419384","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Drone examination has been overall quickly embraced by NDMM (natural disaster mitigation and management) division to survey the state of impacted regions. Manual video analysis by human observers takes time and is subject to mistakes. The human identification examination of pictures caught by drones will give a practical method for saving lives who are being trapped under debris during quakes or in floods and so on. Drone investigation for research and security and search and rescue (SAR) should involve the drone to filter the impacted area using a camera and a model of unmanned area vehicles (UAVs) to identify specific locations where assistance is required. The existing methods (Balmukund et al. 2020) used were faster-region based convolutional neural networks (F-RCNNs), single shot detector (SSD), and region-based fully convolutional network (R-FCN) for the detection of human and recognition of action. Some of the existing methods used 700 images with six classes only, whereas the proposed model uses 1996 images with eight classes. The proposed model is used YOLOv3 (you only look once) algorithm for the detection and recognition of actions. In this study, we provide the fundamental ideas underlying an object detection model. To find the most effective model for human recognition and detection, we trained the YOLOv3 algorithm on the image dataset and evaluated its performance. We compared the outcomes with the existing algorithms like F-RCNN, SSD, and R-FCN. The accuracies of F-RCNN, SSD, R-FCN (existing algorithms), and YOLOv3 (proposed algorithm) are 53%, 73%, 93%, and 94.9%, respectively. Among these algorithms, the YOLOv3 algorithm gives the highest accuracy of 94.9%. The proposed work shows that existing models are inadequate for critical applications like search and rescue, which convinces us to propose a model raised by a pyramidal component extracting SSD in human localization and action recognition. The suggested model is 94.9% accurate when applied to the proposed dataset, which is an important contribution. Likewise, the suggested model succeeds in helping time for expectation in examination with the cutting-edge identification models with existing strategies. The average time taken by our proposed technique to distinguish a picture is 0.40 milisec which is a lot better than the existing method. The proposed model can likewise distinguish video and can be utilized for real-time recognition. The SSD model can likewise use to anticipate messages if present in the picture.
基于YOLOv3算法的灾害搜救人员检测与行动识别
无人机检查已被NDMM(自然灾害缓解和管理)部门全面迅速接受,以调查受影响地区的状况。人工视频分析需要时间,而且容易出错。对无人机拍摄的照片进行人类识别检查,将为在地震或洪水等期间拯救被困在废墟下的生命提供一种实用的方法。用于研究和安全以及搜救(SAR)的无人机调查应涉及无人机使用摄像头和无人驾驶区域车辆(uav)模型过滤受影响区域,以确定需要援助的特定位置。现有的方法(Balmukund et al. 2020)使用更快的基于区域的卷积神经网络(F-RCNNs)、单镜头检测器(SSD)和基于区域的全卷积网络(R-FCN)来检测人类和识别动作。现有的一些方法只使用了包含6个类的700幅图像,而该模型使用了包含8个类的1996幅图像。所提出的模型使用YOLOv3(只看一次)算法来检测和识别动作。在这项研究中,我们提供了目标检测模型的基本思想。为了找到最有效的人类识别和检测模型,我们在图像数据集上训练YOLOv3算法并评估其性能。我们将结果与现有的算法如F-RCNN、SSD和R-FCN进行了比较。现有算法F-RCNN、SSD、R-FCN和YOLOv3的准确率分别为53%、73%、93%和94.9%。在这些算法中,YOLOv3算法的准确率最高,达到94.9%。我们的研究表明,现有的模型在搜索和救援等关键应用中是不够的,这使我们有理由提出一种基于锥体成分提取SSD的人体定位和动作识别模型。该模型在数据集上的准确率为94.9%,这是一个重要的贡献。同样,建议的模型成功地帮助使用现有策略的前沿识别模型检查期望时间。该方法识别图像的平均时间为0.40毫秒,大大优于现有方法。该模型同样可以区分视频,并可用于实时识别。SSD模型同样可以用于预测图片中出现的消息。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信