计算机科学
行人
帧(网络)
图形
人工智能
卷积神经网络
循环神经网络
特征(语言学)
解析
计算机视觉
人工神经网络
理论计算机科学
计算机网络
语言学
哲学
运输工程
工程类
作者
Xingchen Song,Miao Kang,Sanping Zhou,Jianji Wang,Yishu Mao,Nanning Zheng
标识
DOI:10.1109/iros47612.2022.9981690
摘要
Anticipating the future behavior of pedestrians is a crucial part of deploying Automated Driving Systems (ADS) in urban traffic scenarios. Most recent works utilize a convolutional neural network (CNN) to extract visual information, which is then input to a recurrent neural network (RNN) along with pedestrian-specific features like location and speed to obtain temporal features. However, the majority of these approaches lack the ability to parse the relationships of the related objects in the specific traffic scene, which leads to omitting the interactions between the pedestrians and the interactions between the pedestrians and the traffic. For this purpose, we propose a graph-structured model which can dig out pedestrians' dynamic constraints by constructing a traffic-aware scene graph within each frame. In addition, to capture pedestrian movement more effectively, we also introduce a temporal feature representation model, which first uses inter-frame and intra-frame GRU (II-GRU) to mine inter-frame information and intra-frame information together, and then employs a novel attention mechanism to adaptively generate attention weights. Extensive experiments on the JAAD and PIE datasets prove that our proposed model is effective in reaching and enhancing the state-of-the-art performance.
科研通智能强力驱动
Strongly Powered by AbleSci AI