• 场景流与推理辅助的多帧点云目标检测模型

    Multi-frame point cloud object detection with scene flow and inference

    • 随着自动驾驶技术的进步,基于激光雷达点云的3D目标检测已成为关键任务。然而,现有方法大多依赖单帧检测,未充分利用时序信息,导致检测精度不高,存在遮挡和目标丢失等问题。为此,提出了一种新型多帧目标检测模型, 通过集成多帧信息来提高检测性能。为了有效利用多帧信息,所提模型引入了匹配推理模块和场景流模块。其中,匹配推理模块根据目标的初始位置推断其运动方向和速度,更新目标的后续位置。场景流模块则整合了点云中各点的位置信息、速度和方向,为推理过程提供更准确的数据支持。实验表明:所提模型在nuScencs数据集中达到了59.2%的mAP和67.4%的NDS,在KITTI数据集中重要的车类别中等难度评价中取得了89.03%的平均精度,优于大部分目标检测模型。同时,将检测结果可视化,通过多种消融实验进一步验证了该模型的有效性和所提出改进的可行性。

       

      Abstract: With the advancement of autonomous driving technology, 3D object detection based on LiDAR point clouds has become a critical task. However, existing methods mostly rely on single-frame detection, failing to fully utilize temporal information, which leads to lower detection accuracy, occlusion issues, and object loss. a novel multi-frame object detection model that integrates temporal information to improve detection performance is proposed. To effectively leverage multi-frame data, the proposed model introduces a matching inference module and a scene flow module. The matching inference module predicts the object's direction and velocity based on its initial position and updates its subsequent positions. The scene flow module integrates each point’s position, velocity, and direction in the point cloud, providing more accurate data for inference. Experiments show that the proposed model achieves 59.2% mAP and 67.4% NDS on the nuScenes dataset, as well as 89.03% average precision for the car category in the moderate difficulty evaluation on the KITTI dataset, outperforming most object detection models. Additionally, we visualized the detection results and conducted extensive ablation studies, further validating the model's effectiveness and the feasibility of the proposed improvements.

       

    /

    返回文章
    返回