计算机科学
人工智能
卷积神经网络
光学(聚焦)
注意力网络
背景(考古学)
深度学习
RGB颜色模型
分散注意力
人工神经网络
机器学习
计算机视觉
生物
光学
物理
古生物学
神经科学
作者
Rui Fu,Tao Huang,Mingyue Li,Qinyu Sun,Yunxing Chen
标识
DOI:10.1016/j.eswa.2022.119157
摘要
The prediction of the driver’s focus of attention (DFoA) is becoming essential research for the driver distraction detection and intelligent vehicle. Therefore, this work makes an attempt to predict DFoA. However, traffic driving environment is a complex and dynamic changing scene. The existing methods lack full utilization of driving scene information and ignore the importance of different objects or regions of the driving scene. To alleviate this, we propose a multimodal deep neural network based on anthropomorphic attention mechanism and prior knowledge (MDNN-AAM-PK). Specifically, a more comprehensive information of driving scene (RGB images, semantic images, optical flow images and depth images of successive frames) is as the input of MDNN-AAM-PK. An anthropomorphic attention mechanism is developed to calculate the importance of each pixel in the driving scene. A graph attention network is adopted to learn semantic context features. The convolutional long short-term memory network (ConvLSTM) is used to achieve the transition of fused features in successive frames. Furthermore, a training method based on prior knowledge is designed to improve the efficiency of training and the performance of DFoA prediction. These experiments, including experimental comparison with the state-of-the-art methods, the ablation study of the proposed method, the evaluation on different datasets and the visual assessment experiment in vehicle simulation platform, show that the proposed method can accurately predict DFoA and is better than the state-of-the-art methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI