基于双流视频的碰撞和濒临碰撞深度学习模型

IF 7.6 1区 工程技术 Q1 TRANSPORTATION SCIENCE & TECHNOLOGY
{"title":"基于双流视频的碰撞和濒临碰撞深度学习模型","authors":"","doi":"10.1016/j.trc.2024.104794","DOIUrl":null,"url":null,"abstract":"<div><p>The use of videos for effective crash and near-crash prediction can significantly enhance the development of safety countermeasures and emergency response. This paper presents a two-stream hybrid model with temporal and spatial streams for crash and near-crash identification based on front-view video driving data. The novel temporal stream integrates optical flow and TimeSFormer, utilizing divided-space–time attention. The spatial stream employs TimeSFormer with space attention to complement spatial information that is not captured by the optical flow. An XGBoost classifier merges the two streams through late fusion. The proposed approach utilizes data from the Second Strategic Highway Research Program Naturalistic Driving Study, which encompasses 1922 crashes, 6960 near-crashes, and 8611 normal driving segments. The results demonstrate excellent performance, achieving an overall accuracy of 0.894. The F1 scores for crashes, near-crashes, and normal driving segments were 0.760, 0.892, and 0.923, respectively, indicating strong predictive power for all three categories. The proposed approach offers a highly effective and scalable solution for identifying crashes and near-crashes using front-view video driving data and has broad applications in the field of traffic safety.</p></div>","PeriodicalId":54417,"journal":{"name":"Transportation Research Part C-Emerging Technologies","volume":null,"pages":null},"PeriodicalIF":7.6000,"publicationDate":"2024-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Two-stream video-based deep learning model for crashes and near-crashes\",\"authors\":\"\",\"doi\":\"10.1016/j.trc.2024.104794\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>The use of videos for effective crash and near-crash prediction can significantly enhance the development of safety countermeasures and emergency response. This paper presents a two-stream hybrid model with temporal and spatial streams for crash and near-crash identification based on front-view video driving data. The novel temporal stream integrates optical flow and TimeSFormer, utilizing divided-space–time attention. The spatial stream employs TimeSFormer with space attention to complement spatial information that is not captured by the optical flow. An XGBoost classifier merges the two streams through late fusion. The proposed approach utilizes data from the Second Strategic Highway Research Program Naturalistic Driving Study, which encompasses 1922 crashes, 6960 near-crashes, and 8611 normal driving segments. The results demonstrate excellent performance, achieving an overall accuracy of 0.894. The F1 scores for crashes, near-crashes, and normal driving segments were 0.760, 0.892, and 0.923, respectively, indicating strong predictive power for all three categories. The proposed approach offers a highly effective and scalable solution for identifying crashes and near-crashes using front-view video driving data and has broad applications in the field of traffic safety.</p></div>\",\"PeriodicalId\":54417,\"journal\":{\"name\":\"Transportation Research Part C-Emerging Technologies\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":7.6000,\"publicationDate\":\"2024-08-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Transportation Research Part C-Emerging Technologies\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0968090X24003152\",\"RegionNum\":1,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"TRANSPORTATION SCIENCE & TECHNOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Transportation Research Part C-Emerging Technologies","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0968090X24003152","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"TRANSPORTATION SCIENCE & TECHNOLOGY","Score":null,"Total":0}
引用次数: 0

摘要

利用视频进行有效的碰撞和近似碰撞预测可大大促进安全对策的制定和应急响应。本文基于前视驾驶视频数据,提出了一种具有时间流和空间流的双流混合模型,用于碰撞和近碰撞识别。新颖的时间流整合了光流和 TimeSFormer,利用了分时空注意力。空间流采用了带有空间注意力的 TimeSFormer,以补充光学流未捕捉到的空间信息。XGBoost 分类器通过后期融合将两个流合并。所提出的方法利用了第二次公路战略研究计划自然驾驶研究的数据,其中包括 1922 起碰撞事故、6960 起濒临碰撞事故和 8611 个正常驾驶路段。研究结果表明,数据的整体准确度达到 0.894,表现出色。碰撞事故、濒临碰撞事故和正常驾驶路段的 F1 分数分别为 0.760、0.892 和 0.923,表明对所有三个类别都有很强的预测能力。所提出的方法为利用前视驾驶视频数据识别碰撞和近似碰撞提供了一种高效且可扩展的解决方案,在交通安全领域具有广泛的应用前景。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Two-stream video-based deep learning model for crashes and near-crashes

The use of videos for effective crash and near-crash prediction can significantly enhance the development of safety countermeasures and emergency response. This paper presents a two-stream hybrid model with temporal and spatial streams for crash and near-crash identification based on front-view video driving data. The novel temporal stream integrates optical flow and TimeSFormer, utilizing divided-space–time attention. The spatial stream employs TimeSFormer with space attention to complement spatial information that is not captured by the optical flow. An XGBoost classifier merges the two streams through late fusion. The proposed approach utilizes data from the Second Strategic Highway Research Program Naturalistic Driving Study, which encompasses 1922 crashes, 6960 near-crashes, and 8611 normal driving segments. The results demonstrate excellent performance, achieving an overall accuracy of 0.894. The F1 scores for crashes, near-crashes, and normal driving segments were 0.760, 0.892, and 0.923, respectively, indicating strong predictive power for all three categories. The proposed approach offers a highly effective and scalable solution for identifying crashes and near-crashes using front-view video driving data and has broad applications in the field of traffic safety.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
15.80
自引率
12.00%
发文量
332
审稿时长
64 days
期刊介绍: Transportation Research: Part C (TR_C) is dedicated to showcasing high-quality, scholarly research that delves into the development, applications, and implications of transportation systems and emerging technologies. Our focus lies not solely on individual technologies, but rather on their broader implications for the planning, design, operation, control, maintenance, and rehabilitation of transportation systems, services, and components. In essence, the intellectual core of the journal revolves around the transportation aspect rather than the technology itself. We actively encourage the integration of quantitative methods from diverse fields such as operations research, control systems, complex networks, computer science, and artificial intelligence. Join us in exploring the intersection of transportation systems and emerging technologies to drive innovation and progress in the field.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信