论文标题

Pillargrid:3D对象检测的深度学习合作感

PillarGrid: Deep Learning-based Cooperative Perception for 3D Object Detection from Onboard-Roadside LiDAR

论文作者

Bai, Zhengwei, Wu, Guoyuan, Barth, Matthew J., Liu, Yongkang, Sisbot, Emrah Akin, Oguchi, Kentaro

论文摘要

3D对象检测在实现自主驾驶方面起着基本作用,这被认为是从安全,流动性和可持续性的角度来解锁当代运输系统瓶颈的重要关键。从点云中的大多数最先进的(SOTA)对象检测方法是基于单个板上的激光雷达开发的,其性能将不可避免地受到范围和遮挡的限制,尤其是在密集的交通情况下。在本文中,我们提出了一种新型的合作感知方法\ textit {pillargrid},从而融合了来自多个3D激光雷达(板上和路边)的信息,以提高连接和自动化车辆(CAVS)的情况意识。 Pillargrid由四个主要阶段组成:1)点云的合作预处理,2)支柱的脱氧化和特征提取,3)来自多个传感器的特征的网格深融合,以及4)卷积神经网络(CNN)基于增强的3D对象检测。开发了一个新型的合作感知平台,用于模型培训和测试。广泛的实验表明,Pillargrid的表现优于SOTA基于SOTA的3D对象检测方法,相对于准确性和范围较大。

3D object detection plays a fundamental role in enabling autonomous driving, which is regarded as the significant key to unlocking the bottleneck of contemporary transportation systems from the perspectives of safety, mobility, and sustainability. Most of the state-of-the-art (SOTA) object detection methods from point clouds are developed based on a single onboard LiDAR, whose performance will be inevitably limited by the range and occlusion, especially in dense traffic scenarios. In this paper, we propose \textit{PillarGrid}, a novel cooperative perception method fusing information from multiple 3D LiDARs (both on-board and roadside), to enhance the situation awareness for connected and automated vehicles (CAVs). PillarGrid consists of four main phases: 1) cooperative preprocessing of point clouds, 2) pillar-wise voxelization and feature extraction, 3) grid-wise deep fusion of features from multiple sensors, and 4) convolutional neural network (CNN)-based augmented 3D object detection. A novel cooperative perception platform is developed for model training and testing. Extensive experimentation shows that PillarGrid outperforms the SOTA single-LiDAR-based 3D object detection methods with respect to both accuracy and range by a large margin.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源