DrivAid: Augmenting Driving Analytics with Multi-Modal Information

被引:0
|
作者
Qi, Bozhao [1 ]
Liu, Peng [1 ]
Ji, Tao [1 ]
Zhao, Wei [1 ]
Banerjee, Suman [1 ]
机构
[1] Univ Wisconsin, Dept Comp Sci, 1210 W Dayton St, Madison, WI 53706 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The way people drive vehicles has a great impact on traffic safety, fuel consumption, and passenger experience. Many research and commercial efforts today have primarily leveraged the Inertial Measurement Unit (IMU) to characterize, profile, and understand how well people drive their vehicles. In this paper, we observe that such IMU data alone cannot always reveal a driver's context and therefore does not provide a comprehensive understanding of a driver's actions. We believe that an audiovisual infrastructure, with cameras and microphones, can be well leveraged to augment IMU data to reveal driver context and improve analytics. For instance, such an audio-visual system can easily discern whether a hard braking incident, as detected by an accelerometer, is the result of inattentive driving (e.g., a distracted driver) or evidence of alertness (e.g., a driver avoids a deer). The focus of this work has been to design a relatively lowcost audio-visual infrastructure through which it is practical to gather such context information from various sensors and to develop a comprehensive understanding of why a particular driver may have taken different actions. In particular, we build a system called DrivAid, that collects and analyzes visual and audio signals in real time with computer vision techniques on a vehicle-based edge computing platform, to complement the signals from traditional motion sensors. Driver privacy is preserved since the audio-visual data is mainly processed locally. We implement DrivAid on a low-cost embedded computer with GPU and high-performance deep learning inference support. In total, we have collected more than 1550 miles of driving data from multiple vehicles to build and test our system. The evaluation results show that DrivAid is able to process video streams from 4 cameras at a rate of 10 frames per second. DrivAid can achieve an average of 90% event detection accuracy and provide reasonable evaluation feedbacks to users in real time. With the efficient design, for a single trip, only around 36% of audio-visual data needs to be analyzed on average.
引用
收藏
页数:8
相关论文
共 50 条
  • [41] A Stumble Detection Method for Programming with Multi-modal Information
    Oka, Hiroki
    Ohnishi, Ayumi
    Terada, Tsutomu
    Tsukamoto, Masahiko
    [J]. ADVANCES IN MOBILE COMPUTING AND MULTIMEDIA INTELLIGENCE, MOMM 2022, 2022, 13634 : 169 - 174
  • [42] Facial emotion recognition using multi-modal information
    De Silva, LC
    Miyasato, T
    Nakatsu, R
    [J]. ICICS - PROCEEDINGS OF 1997 INTERNATIONAL CONFERENCE ON INFORMATION, COMMUNICATIONS AND SIGNAL PROCESSING, VOLS 1-3: THEME: TRENDS IN INFORMATION SYSTEMS ENGINEERING AND WIRELESS MULTIMEDIA COMMUNICATIONS, 1997, : 397 - 401
  • [43] Multi-Modal Information Fusion for Localization of Emergency Vehicles
    Joshi, Aruna Kumar
    Kulkarni, Shrinivasrao B.
    [J]. INTERNATIONAL JOURNAL OF IMAGE AND GRAPHICS, 2024,
  • [44] Using information gain to improve multi-modal information retrieval systems
    Martin-Valdivia, M. T.
    Diaz-Galiano, M. C.
    Montejo-Raez, A.
    Urena-Lopez, L. A.
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2008, 44 (03) : 1146 - 1158
  • [45] Multi-modal information integration for interactive multi-agent systems
    Yamaguchi, T
    Sato, M
    Takagi, T
    [J]. JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 1998, 23 (2-4) : 183 - 199
  • [46] Multi-modal Information Integration for Interactive Multi-agent Systems
    Toru Yamaguchi
    Makoto Sato
    Tomohiro Takagi
    [J]. Journal of Intelligent and Robotic Systems, 1998, 23 : 183 - 199
  • [47] Multi-modal object detection and localization for high integrity driving assistance
    Sergio Alberto Rodríguez Flórez
    Vincent Frémont
    Philippe Bonnifait
    Véronique Cherfaoui
    [J]. Machine Vision and Applications, 2014, 25 : 583 - 598
  • [48] Leveraging Uncertainties for Deep Multi-modal Object Detection in Autonomous Driving
    Feng, Di
    Cao, Yifan
    Rosenbaum, Lars
    Timm, Fabian
    Dietmayer, Klaus
    [J]. 2020 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2020, : 871 - 878
  • [49] Multi-modal object detection and localization for high integrity driving assistance
    Florez, Sergio Alberto Rodriguez
    Fremont, Vincent
    Bonnifait, Philippe
    Cherfaoui, Veronique
    [J]. MACHINE VISION AND APPLICATIONS, 2014, 25 (03) : 583 - 598
  • [50] Multi-modal policy fusion for end-to-end autonomous driving
    Huang, Zhenbo
    Sun, Shiliang
    Zhao, Jing
    Mao, Liang
    [J]. INFORMATION FUSION, 2023, 98