点云
计算机科学
目标检测
人工智能
推论
模式识别(心理学)
特征提取
对象(语法)
特征(语言学)
激光雷达
职位(财务)
计算机视觉
支柱
点(几何)
数据挖掘
数学
工程类
遥感
哲学
语言学
几何学
结构工程
财务
经济
地质学
作者
Ming-Jen Chang,Chih–Jen Cheng,Ching-Chun Hsiao,I-Fan Chou,Ching-Chun Huang
标识
DOI:10.1109/avss56176.2022.9959572
摘要
Although pillar-based 3D object detection methods can balance the performance and inference speed, the inconsistent object features caused by dramatic sparsity drops of LiDAR point clouds sabotage the detection accuracy. We present a novel and efficient plug-in method, SVDnet, to improve the state-of-the-art pillar-based models. First, a novel low-rank objective loss is introduced to extract distance-aware vehicle features and suppress the other variations. Next, we alleviated the remaining feature inconsistency caused by object positions with two strategies. One is a Distance Alignment Ratio-generation Network (DARN), which fuses multi-scale features by distance-adaptive ratios. The other is a position attention network that modulates features based on positions. Our results on the KITTI dataset show that SVDnet improves the pillar methods and outperforms the other plug-in strategies in accuracy and speed.
科研通智能强力驱动
Strongly Powered by AbleSci AI