DrivAid: Augmenting Driving Analytics with Multi-Modal Information

被引:0
|
作者
Qi, Bozhao [1 ]
Liu, Peng [1 ]
Ji, Tao [1 ]
Zhao, Wei [1 ]
Banerjee, Suman [1 ]
机构
[1] Univ Wisconsin, Dept Comp Sci, 1210 W Dayton St, Madison, WI 53706 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The way people drive vehicles has a great impact on traffic safety, fuel consumption, and passenger experience. Many research and commercial efforts today have primarily leveraged the Inertial Measurement Unit (IMU) to characterize, profile, and understand how well people drive their vehicles. In this paper, we observe that such IMU data alone cannot always reveal a driver's context and therefore does not provide a comprehensive understanding of a driver's actions. We believe that an audiovisual infrastructure, with cameras and microphones, can be well leveraged to augment IMU data to reveal driver context and improve analytics. For instance, such an audio-visual system can easily discern whether a hard braking incident, as detected by an accelerometer, is the result of inattentive driving (e.g., a distracted driver) or evidence of alertness (e.g., a driver avoids a deer). The focus of this work has been to design a relatively lowcost audio-visual infrastructure through which it is practical to gather such context information from various sensors and to develop a comprehensive understanding of why a particular driver may have taken different actions. In particular, we build a system called DrivAid, that collects and analyzes visual and audio signals in real time with computer vision techniques on a vehicle-based edge computing platform, to complement the signals from traditional motion sensors. Driver privacy is preserved since the audio-visual data is mainly processed locally. We implement DrivAid on a low-cost embedded computer with GPU and high-performance deep learning inference support. In total, we have collected more than 1550 miles of driving data from multiple vehicles to build and test our system. The evaluation results show that DrivAid is able to process video streams from 4 cameras at a rate of 10 frames per second. DrivAid can achieve an average of 90% event detection accuracy and provide reasonable evaluation feedbacks to users in real time. With the efficient design, for a single trip, only around 36% of audio-visual data needs to be analyzed on average.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Special issue on multi-modal information learning and analytics on big data
    Xiaomeng Ma
    Yan Sun
    [J]. Neural Computing and Applications, 2022, 34 : 3299 - 3300
  • [2] Special issue on multi-modal information learning and analytics for smart city
    Zheng Xu
    Qingyuan Zhou
    [J]. Journal of Ambient Intelligence and Humanized Computing, 2020, 11 : 3471 - 3472
  • [3] Special issue on multi-modal information learning and analytics for smart city
    Xu, Zheng
    Zhou, Qingyuan
    [J]. JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2020, 11 (09) : 3471 - 3472
  • [4] Special issue on multi-modal information learning and analytics on big data
    Ma, Xiaomeng
    Sun, Yan
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (05): : 3299 - 3300
  • [5] Editorial: Special Issue on Multi-modal Information mining and Analytics for Environmental Technology & Innovation
    Xu, Zheng
    Yen, Neil
    Sugumaran, Vijayan
    [J]. ENVIRONMENTAL TECHNOLOGY & INNOVATION, 2022, 28
  • [6] Multi-modal Experts Network for Autonomous Driving
    Fang, Shihong
    Choromanska, Anna
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 6439 - 6445
  • [7] Multi-modal Contrastive Learning for Healthcare Data Analytics
    Li, Rui
    Gao, Jing
    [J]. 2022 IEEE 10TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI 2022), 2022, : 120 - 127
  • [8] Pervasive Physical Analytics using Multi-Modal Sensing
    Sen, Sougata
    [J]. 2016 8TH INTERNATIONAL CONFERENCE ON COMMUNICATION SYSTEMS AND NETWORKS (COMSNETS), 2016,
  • [9] MULTI-MODAL TRAVEL INFORMATION ON THE WEB
    Pun-Cheng, Lilian S. C.
    Shea, Geoffrey Y. K.
    Mok, Esmond C. M.
    [J]. TRANSPORTATION AND LOGISTICS, 2003, : 285 - 290
  • [10] A Train Driver Fatigue Driving Detection Method Based on Multi-modal Information Fusion
    Li, Xiaoping
    Bai, Chao
    [J]. Tiedao Xuebao/Journal of the China Railway Society, 2022, 44 (06): : 56 - 65