DrivAid: Augmenting Driving Analytics with Multi-Modal Information

Bozhao Qi, Peng Liu, Tao Ji, Wei Zhao, Suman Banerjee
{"title":"DrivAid: Augmenting Driving Analytics with Multi-Modal Information","authors":"Bozhao Qi, Peng Liu, Tao Ji, Wei Zhao, Suman Banerjee","doi":"10.1109/VNC.2018.8628415","DOIUrl":null,"url":null,"abstract":"The way people drive vehicles has a great impact on traffic safety, fuel consumption, and passenger experience. Many research and commercial efforts today have primarily leveraged the Inertial Measurement Unit (IMU) to characterize, profile, and understand how well people drive their vehicles. In this paper, we observe that such IMU data alone cannot always reveal a driver’s context and therefore does not provide a comprehensive understanding of a driver’s actions. We believe that an audio-visual infrastructure, with cameras and microphones, can be well leveraged to augment IMU data to reveal driver context and improve analytics. For instance, such an audio-visual system can easily discern whether a hard braking incident, as detected by an accelerometer, is the result of inattentive driving (e.g., a distracted driver) or evidence of alertness (e.g., a driver avoids a deer).The focus of this work has been to design a relatively low-cost audio-visual infrastructure through which it is practical to gather such context information from various sensors and to develop a comprehensive understanding of why a particular driver may have taken different actions. In particular, we build a system called DrivAid, that collects and analyzes visual and audio signals in real time with computer vision techniques on a vehicle-based edge computing platform, to complement the signals from traditional motion sensors. Driver privacy is preserved since the audio-visual data is mainly processed locally. We implement DrivAid on a low-cost embedded computer with GPU and high-performance deep learning inference support. In total, we have collected more than 1550 miles of driving data from multiple vehicles to build and test our system. The evaluation results show that DrivAid is able to process video streams from 4 cameras at a rate of 10 frames per second. DrivAid can achieve an average of 90% event detection accuracy and provide reasonable evaluation feedbacks to users in real time. With the efficient design, for a single trip, only around 36% of audio-visual data needs to be analyzed on average.","PeriodicalId":335017,"journal":{"name":"2018 IEEE Vehicular Networking Conference (VNC)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"26","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE Vehicular Networking Conference (VNC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/VNC.2018.8628415","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 26

Abstract

The way people drive vehicles has a great impact on traffic safety, fuel consumption, and passenger experience. Many research and commercial efforts today have primarily leveraged the Inertial Measurement Unit (IMU) to characterize, profile, and understand how well people drive their vehicles. In this paper, we observe that such IMU data alone cannot always reveal a driver’s context and therefore does not provide a comprehensive understanding of a driver’s actions. We believe that an audio-visual infrastructure, with cameras and microphones, can be well leveraged to augment IMU data to reveal driver context and improve analytics. For instance, such an audio-visual system can easily discern whether a hard braking incident, as detected by an accelerometer, is the result of inattentive driving (e.g., a distracted driver) or evidence of alertness (e.g., a driver avoids a deer).The focus of this work has been to design a relatively low-cost audio-visual infrastructure through which it is practical to gather such context information from various sensors and to develop a comprehensive understanding of why a particular driver may have taken different actions. In particular, we build a system called DrivAid, that collects and analyzes visual and audio signals in real time with computer vision techniques on a vehicle-based edge computing platform, to complement the signals from traditional motion sensors. Driver privacy is preserved since the audio-visual data is mainly processed locally. We implement DrivAid on a low-cost embedded computer with GPU and high-performance deep learning inference support. In total, we have collected more than 1550 miles of driving data from multiple vehicles to build and test our system. The evaluation results show that DrivAid is able to process video streams from 4 cameras at a rate of 10 frames per second. DrivAid can achieve an average of 90% event detection accuracy and provide reasonable evaluation feedbacks to users in real time. With the efficient design, for a single trip, only around 36% of audio-visual data needs to be analyzed on average.
DrivAid:使用多模态信息增强驾驶分析
人们驾驶车辆的方式对交通安全、燃油消耗和乘客体验有很大的影响。今天,许多研究和商业努力主要利用惯性测量单元(IMU)来表征、描述和了解人们驾驶车辆的情况。在本文中,我们观察到这样的IMU数据本身并不能总是揭示驾驶员的上下文,因此不能提供对驾驶员行为的全面理解。我们相信,配备摄像头和麦克风的视听基础设施可以很好地增强IMU数据,以揭示驾驶员的背景并改进分析。例如,这种视听系统可以很容易地分辨出加速计检测到的硬制动事件是由于驾驶疏忽(例如,驾驶员分心)还是警觉(例如,驾驶员避开鹿)造成的。这项工作的重点是设计一种相对低成本的视听基础设施,通过这种基础设施,可以从各种传感器收集此类背景信息,并全面了解特定驾驶员可能采取不同行动的原因。特别是,我们建立了一个名为DrivAid的系统,该系统通过基于车辆边缘计算平台的计算机视觉技术实时收集和分析视觉和音频信号,以补充传统运动传感器的信号。由于视听数据主要在本地处理,因此保护了驾驶员的隐私。我们在具有GPU和高性能深度学习推理支持的低成本嵌入式计算机上实现了DrivAid。为了构建和测试我们的系统,我们总共从多辆车那里收集了超过1550英里的驾驶数据。评估结果表明,DrivAid能够以每秒10帧的速率处理来自4个摄像头的视频流。DrivAid可以实现平均90%的事件检测准确率,并实时向用户提供合理的评价反馈。通过高效的设计,单次行程平均只需要分析约36%的视听数据。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信