A. Agrawal, Kadamb Agarwal, J. Choudhary, Aradhita Bhattacharya, Srihitha Tangudu, Nishkarsh Makhija, R. B
{"title":"Automatic Traffic Accident Detection System Using ResNet and SVM","authors":"A. Agrawal, Kadamb Agarwal, J. Choudhary, Aradhita Bhattacharya, Srihitha Tangudu, Nishkarsh Makhija, R. B","doi":"10.1109/ICRCICN50933.2020.9296156","DOIUrl":null,"url":null,"abstract":"The rate of road accidents has increased to a large extent over the last few years. This has eventually resulted in a huge loss of lives and property. Hence, the need of the hour has aroused to detect such accident spots as quickly as possible so that proper life-saving measures can be taken and the mishap prone areas can be put on alert. In order to address this problem, we propose a Machine Learning and Deep Learning based model on the concepts of Clustering and Classification that can be used to detect accidents from the traffic surveillance cameras. Firstly all the videos are split up into smaller shots according to scene changes. And then key frames are extracted from each shot based on histogram difference of consecutive frames. Then distance between the vehicles is determined to detect the potential accident. The obtained key frames are passed through a ResNet50 architecture for feature extraction. After obtaining the feature vectors of all videos, K-Means clustering has been applied to obtain Bag of Visual Words(BOVW). Finally, these bag of visual words is sent as input to a Support Vector Machine(SVM) classifier that outputs if a video contained an accident or not. The proposed method has an accuracy of 94.14%.","PeriodicalId":138966,"journal":{"name":"2020 Fifth International Conference on Research in Computational Intelligence and Communication Networks (ICRCICN)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 Fifth International Conference on Research in Computational Intelligence and Communication Networks (ICRCICN)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICRCICN50933.2020.9296156","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
The rate of road accidents has increased to a large extent over the last few years. This has eventually resulted in a huge loss of lives and property. Hence, the need of the hour has aroused to detect such accident spots as quickly as possible so that proper life-saving measures can be taken and the mishap prone areas can be put on alert. In order to address this problem, we propose a Machine Learning and Deep Learning based model on the concepts of Clustering and Classification that can be used to detect accidents from the traffic surveillance cameras. Firstly all the videos are split up into smaller shots according to scene changes. And then key frames are extracted from each shot based on histogram difference of consecutive frames. Then distance between the vehicles is determined to detect the potential accident. The obtained key frames are passed through a ResNet50 architecture for feature extraction. After obtaining the feature vectors of all videos, K-Means clustering has been applied to obtain Bag of Visual Words(BOVW). Finally, these bag of visual words is sent as input to a Support Vector Machine(SVM) classifier that outputs if a video contained an accident or not. The proposed method has an accuracy of 94.14%.
在过去的几年里,交通事故的发生率在很大程度上增加了。这最终造成了巨大的生命和财产损失。因此,迫切需要尽快发现这些事故地点,以便采取适当的救生措施,并提高事故易发地区的警戒水平。为了解决这个问题,我们提出了一个基于机器学习和深度学习的模型,该模型基于聚类和分类的概念,可用于从交通监控摄像机中检测事故。首先,所有的视频根据场景的变化被分割成更小的镜头。然后根据连续帧的直方图差异从每个镜头中提取关键帧。然后确定车辆之间的距离,以检测潜在的事故。获得的关键帧通过ResNet50架构进行特征提取。在获得所有视频的特征向量后,应用K-Means聚类得到视觉词包(Bag of Visual Words, BOVW)。最后,这些视觉词包作为输入发送给支持向量机(SVM)分类器,该分类器输出视频是否包含事故。该方法的准确率为94.14%。