J. Evangelin, Deva Sheela, P. Arockia, J. Rani, M. A. Paul
{"title":"无人机视频中基于超像素传输图的深度神经网络目标检测","authors":"J. Evangelin, Deva Sheela, P. Arockia, J. Rani, M. A. Paul","doi":"10.1080/13682199.2023.2195121","DOIUrl":null,"url":null,"abstract":"ABSTRACT Object detection has become a very prominent subject for research in recent times. This study's main goal is to suggest a technique for video saliency object detection. It seems to sense that using the depth information in photos to detect salient things. Since depth offers abundant information about scene structure, object forms, and other 3D cues. This information is very compatible to distinguish between objects in the foreground and background. As a result of the high object density, small object size, and cluttered background, aerial photos and movies provide results with low precision. In this paper, the proposed SPTM (Super Pixel Transmission Map)-YOLO model, the input RGB image has applied Dark Channel Prior (DCP) method for estimating the transmission map. From the transmission map only, the background probability is estimated with the help of SLIC (simple linear iterative clustering algorithm) superpixel segmentation. That foreground extracted image is further learned with YOLO architecture to detect the objects effectively. For object detection in aerial images, this proposed SPTM-YOLO approach outperforms classic YOLO by up to 6% accuracy. Accurate detection of things that are small in size, partially occluded, and out of view is possible.","PeriodicalId":22456,"journal":{"name":"The Imaging Science Journal","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2023-04-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Super pixels transmission map-based object detection using deep neural network in UAV video\",\"authors\":\"J. Evangelin, Deva Sheela, P. Arockia, J. Rani, M. A. Paul\",\"doi\":\"10.1080/13682199.2023.2195121\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"ABSTRACT Object detection has become a very prominent subject for research in recent times. This study's main goal is to suggest a technique for video saliency object detection. It seems to sense that using the depth information in photos to detect salient things. Since depth offers abundant information about scene structure, object forms, and other 3D cues. This information is very compatible to distinguish between objects in the foreground and background. As a result of the high object density, small object size, and cluttered background, aerial photos and movies provide results with low precision. In this paper, the proposed SPTM (Super Pixel Transmission Map)-YOLO model, the input RGB image has applied Dark Channel Prior (DCP) method for estimating the transmission map. From the transmission map only, the background probability is estimated with the help of SLIC (simple linear iterative clustering algorithm) superpixel segmentation. That foreground extracted image is further learned with YOLO architecture to detect the objects effectively. For object detection in aerial images, this proposed SPTM-YOLO approach outperforms classic YOLO by up to 6% accuracy. Accurate detection of things that are small in size, partially occluded, and out of view is possible.\",\"PeriodicalId\":22456,\"journal\":{\"name\":\"The Imaging Science Journal\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-04-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"The Imaging Science Journal\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1080/13682199.2023.2195121\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Imaging Science Journal","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1080/13682199.2023.2195121","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Super pixels transmission map-based object detection using deep neural network in UAV video
ABSTRACT Object detection has become a very prominent subject for research in recent times. This study's main goal is to suggest a technique for video saliency object detection. It seems to sense that using the depth information in photos to detect salient things. Since depth offers abundant information about scene structure, object forms, and other 3D cues. This information is very compatible to distinguish between objects in the foreground and background. As a result of the high object density, small object size, and cluttered background, aerial photos and movies provide results with low precision. In this paper, the proposed SPTM (Super Pixel Transmission Map)-YOLO model, the input RGB image has applied Dark Channel Prior (DCP) method for estimating the transmission map. From the transmission map only, the background probability is estimated with the help of SLIC (simple linear iterative clustering algorithm) superpixel segmentation. That foreground extracted image is further learned with YOLO architecture to detect the objects effectively. For object detection in aerial images, this proposed SPTM-YOLO approach outperforms classic YOLO by up to 6% accuracy. Accurate detection of things that are small in size, partially occluded, and out of view is possible.