计算机科学
人工智能
点云
计算机视觉
目标检测
最小边界框
对象(语法)
特征(语言学)
像素
交叉口(航空)
模式识别(心理学)
图像(数学)
工程类
语言学
哲学
航空航天工程
作者
Rui Hou,Guangzhu Chen,Yinhe Han,Zaizuo Tang,Qingjun Ru
标识
DOI:10.1016/j.asoc.2021.108245
摘要
3D object detection technology is of great significance to realize intelligent perception and ensure the production safety of a workshop. Existing 3D object detection relies on large-scale, high-quality 3D annotation data and is unsuitable for actual workshop scenes' perception. This paper proposes a multi-modal feature fusion 3D object detection method (MFF3D) for a production workshop. The design of MFF3D includes the following steps: (1) Improved YOLOv3 attains the 2D prior region of an object, and RGB-D saliency detection obtains the object image pixels in that region. (2) Depth image pixels corresponding to the object are projected to generate the object's frustum point cloud, and a multi-modal feature fusion strategy simplifies the object's frustum point cloud, so as to remove outlier points and reduce the number of point clouds. This can replace the 3D object reasoning process based on deep neural networks; (3) An axis-aligned bounding box algorithm is used to generate the object's 3D bounding box, and principal component analysis algorithm (PCA) is used to calculate the object's pose information. MFF3D is applied in the workshop, and experiments verify the feasibility and detection accuracy. We set up a production workshop object dataset (PWOD) for experimental evaluation. In the case of a small amount of 2D annotation data and no 3D annotation data, experimental results show that when the threshold value of intersection over union of 3D object (IoU3D) is 0.5, the mean average precision value of 3D object (mAP3D) reaches 60.31, and the detection speed reaches 3 FPS. MFF3D does not rely on 3D annotation data and can effectively detect objects of a production workshop.
科研通智能强力驱动
Strongly Powered by AbleSci AI