R. Kayalvizhi, S. Malarvizhi, S. Choudhury, A. Topkar, P. Vijayakumar
{"title":"Detection of sharp objects using deep neural network based object detection algorithm","authors":"R. Kayalvizhi, S. Malarvizhi, S. Choudhury, A. Topkar, P. Vijayakumar","doi":"10.1109/ICCCSP49186.2020.9315272","DOIUrl":null,"url":null,"abstract":"Deep learning algorithms have the ability to learn complex functions and provide state-of-the-art results for com-puter vision problems. In recent times, these algorithms far exceeded the existing computer vision based techniques for object detection in X-ray imaging systems. So far, in literature single class of object namely gun and its parts were considered for detection using the SIXray10 database. We propose deep learning-based solution for the detection of sharp objects namely knife, scissors, wrench, pliers in the SIXray 10database. We propose two models namely model A and model B using a common object detection algorithm- YOLOv3 (You Only Look Once) with InceptionV3 and ResNet-50. YOLO is a deep neural network based object detection algorithm that performs the task in one-shot which allows real time inference in video of 15-30 fps. The model is FCN (Fully Convolutional Network) as has the capacity to perform both regression and classification by sharing weights for both the tasks. The network predicts a rectangular box called bounding box around the predicted object of interest along with the associated class. We analyze the performance of both model in terms of mAP. We achieve mean accuracy of 59.95% for model-A and 63.35% for Model-B. The most daunting part of the project is the low ratio of harmful to nonharmful items. By performing rigorous experiments we came up with the best set of possible results which uses varied pretrained neural networks for feature extraction in tandem with YOLO model for object detection. We endeavor to improve on these existing results so as these systems can be successfully deployed in airports to minimize human error and improve security in such environments.","PeriodicalId":310458,"journal":{"name":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 4th International Conference on Computer, Communication and Signal Processing (ICCCSP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCCSP49186.2020.9315272","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Deep learning algorithms have the ability to learn complex functions and provide state-of-the-art results for com-puter vision problems. In recent times, these algorithms far exceeded the existing computer vision based techniques for object detection in X-ray imaging systems. So far, in literature single class of object namely gun and its parts were considered for detection using the SIXray10 database. We propose deep learning-based solution for the detection of sharp objects namely knife, scissors, wrench, pliers in the SIXray 10database. We propose two models namely model A and model B using a common object detection algorithm- YOLOv3 (You Only Look Once) with InceptionV3 and ResNet-50. YOLO is a deep neural network based object detection algorithm that performs the task in one-shot which allows real time inference in video of 15-30 fps. The model is FCN (Fully Convolutional Network) as has the capacity to perform both regression and classification by sharing weights for both the tasks. The network predicts a rectangular box called bounding box around the predicted object of interest along with the associated class. We analyze the performance of both model in terms of mAP. We achieve mean accuracy of 59.95% for model-A and 63.35% for Model-B. The most daunting part of the project is the low ratio of harmful to nonharmful items. By performing rigorous experiments we came up with the best set of possible results which uses varied pretrained neural networks for feature extraction in tandem with YOLO model for object detection. We endeavor to improve on these existing results so as these systems can be successfully deployed in airports to minimize human error and improve security in such environments.