计算机科学
计算机视觉
人工智能
点云
过度拟合
视频跟踪
对象(语法)
特征(语言学)
跟踪(教育)
目标检测
利用
云计算
点(几何)
光流
编码(集合论)
图像(数学)
模式识别(心理学)
人工神经网络
教育学
几何学
心理学
程序设计语言
集合(抽象数据类型)
计算机安全
哲学
操作系统
语言学
数学
作者
Yanding Yang,Kun Jiang,Diange Yang,Yanqin Jiang,Xiaowei Lu
出处
期刊:IEEE Signal Processing Letters
[Institute of Electrical and Electronics Engineers]
日期:2022-01-01
卷期号:29: 1579-1583
被引量:4
标识
DOI:10.1109/lsp.2022.3185948
摘要
Non-visual range sensors such as Lidar have shown the potential to detect, locate and track objects in complex dynamic scenes thanks to their higher stability in comparison with vision-based sensors like cameras. However, due to the disorder, sparsity, and irregularity of the point cloud, it is much more challenging to take advantage of the temporal information in the dynamic 3D point cloud sequences, as it has been done in the image sequences for improving detection and tracking. In this paper, we propose a novel scene-flow-based point cloud feature fusion module to tackle this challenge, based on which a 3D object tracking framework is also achieved to exploit the temporal motion information. Moreover, we carefully designed several training schemes that contribute to the success of this new module by eliminating the issues of overfitting and long-tailed distribution of object categories. Extensive experiments on the public KITTI 3D object tracking dataset demonstrate the effectiveness of the proposed method by achieving superior results to the baselines. The source code is available at https://github.com/Tsinghua-OpenICV/SharingVan-OpenPCDet.
科研通智能强力驱动
Strongly Powered by AbleSci AI