作者投稿和查稿 主编审稿 专家审稿 编委审稿 远程编辑

计算机工程 ›› 2023, Vol. 49 ›› Issue (11): 238-246. doi: 10.19678/j.issn.1000-3428.0066214

• 图形图像处理 • 上一篇    下一篇

基于多方向特征融合的室外三维目标检测方法

雷嘉铭1,2, 俞辉2,3,*, 夏羽4, 郭杰龙2,3, 魏宪2,3   

  1. 1. 福州大学 先进制造学院, 福建 泉州 362000
    2. 中国科学院海西研究院 泉州装备制造研究中心, 福建 泉州 362000
    3. 中国科学院福建物质结构研究所, 福州 350108
    4. 上海宇航系统工程研究所, 上海 200000
  • 收稿日期:2022-11-09 出版日期:2023-11-15 发布日期:2023-11-08
  • 通讯作者: 俞辉
  • 作者简介:

    雷嘉铭(1995—),男,硕士研究生,主研方向为三维目标检测、机器视觉

    夏羽,工程师、硕士

    郭杰龙,工程师、硕士

    魏宪,研究员、博士

  • 基金资助:
    福建省科技计划项目(2021T3003); 泉州市科技项目(2021C065L)

Outdoor 3D Object Detection Method Based on Multi-Direction Features Fusion

Jiaming LEI1,2, Hui YU2,3,*, Yu XIA4, Jielong GUO2,3, Xian WEI2,3   

  1. 1. School of Advanced Manufacturing, Fuzhou University, Quanzhou 362000, Fujian, China
    2. Quanzhou Institute of Equipment Manufacturing Haixi Institutes, Chinese Academy of Sciences, Quanzhou 362000, Fujian, China
    3. Fujian Institute of Research on the Structure of Matter, Chinese Academy of Sciences, Fuzhou 350108, China
    4. Shanghai Institute of Aerospace System Engineering, Shanghai 200000, China
  • Received:2022-11-09 Online:2023-11-15 Published:2023-11-08
  • Contact: Hui YU

摘要:

三维目标检测方法是自动驾驶环境感知系统的重要技术之一,但是现有三维目标检测方法大多存在目标位置识别精度不足、目标朝向预测偏差较大的问题。为此,提出一种基于多方向特征融合的三维目标检测方法。对场景点云进行数据编码,建模距离、夹角等信息并转换成伪图像。利用所提多方向特征融合骨干网络进行特征提取和融合,基于多方向融合特征,利用以中心为基准的检测器进行潜在目标的回归和预测。该方法通过点间距离、夹角建模提供点间的相互关系,丰富特征信息,利用多方向特征融合增强骨干网络的特征提取能力,从而提高位置及朝向的预测精度。实验结果表明,该方法在KITTI和nuScenes数据集上的平均精度均值分别为64.28%和50.17%,相比次优方法分别提升了0.36和1.30个百分点,且在朝向预测准确率对比实验中均取得了最好的平均朝向相似度和平均朝向误差。此外,泛化性实验结果验证了所提多方向特征融合骨干网络能提高网络检测能力并大幅减少参数量,从而提升检测方法的应用表现。

关键词: 机器视觉, 激光雷达, 三维目标检测, 自动驾驶, 点云

Abstract:

The 3D object detection method is one of the significant technologies in the environmental perception of autonomous driving. However, most existing 3D object detection methods have the problem of inaccurate position recognition and large orientation deviation. To address these issues, a 3D object detection method based on multi-direction features fusion is proposed. First, to perform data encoding for a point cloud scenario, modeling distance and angle and transforming into pseudo image. Second, a multi-direction feature-fusion backbone is proposed for features extraction and fusion. Finally, based on the fused features, a center-based detector regresses and predicts potential objects. Distance-angle modeling can supply the relationship between points and enrich features. The multi-direction feature-fusion backbone enhances the ability of features extraction and improves the accuracy of position and orientation estimation. The experimental results show that the mean Average Precision(mAP)of this method on the KITTI and nuScenes datasets was 64.28% and 50.17%, respectively, which is an improvement of 0.36 and 1.30 percentage points, respectively, compared to those of the suboptimal method. In addition, the best Average Orientation Similarity(AOS)and mean Average Orientation Error(mAOE)were achieved in the orientation prediction accuracy comparison experiments. The generalization experimental results verified that the proposed multi-direction feature-fusion backbone network can improve network detection ability and significantly reduce the number of parameters, thereby improving the application performance of the detection method.

Key words: machine vision, lidar, 3D object detection, autonomous driving, point cloud