4.7 Article

Object detection using depth completion and camera-LiDAR fusion for autonomous driving

期刊

INTEGRATED COMPUTER-AIDED ENGINEERING
卷 29, 期 3, 页码 241-258

出版社

IOS PRESS
DOI: 10.3233/ICA-220681

关键词

Autonomous driving; data fusion; deep learning; object detection; transfer learning

资金

  1. FEDER/Ministerio de Ciencia e Innovacion - Agencia Estatal de Investigacion/Proyecto [PID2020-117954RB-C22]
  2. Andalusian Regional Government [P18-RT-2778, US-1263341]
  3. [FPU18/00622]

向作者/读者索取更多资源

This paper proposes a novel data fusion architecture for object detection in autonomous driving, using camera and LiDAR data to achieve reliable performance. With deep learning models and sensor data, our approach significantly outperforms previous methods in various weather and lighting conditions.
Autonomous vehicles are equipped with complimentary sensors to perceive the environment accurately. Deep learning models have proven to be the most effective approach for computer vision problems. Therefore, in autonomous driving, it is essential to design reliable networks to fuse data from different sensors. In this work, we develop a novel data fusion architecture using camera and LiDAR data for object detection in autonomous driving. Given the sparsity of LiDAR data, developing multimodal fusion models is a challenging task. Our proposal integrates an efficient LiDAR sparse-to-dense completion network into the pipeline of object detection models, achieving a more robust performance at different times of the day. The Waymo Open Dataset has been used for the experimental study, which is the most diverse detection benchmark in terms of weather and lighting conditions. The depth completion network is trained with the KITTI depth dataset, and transfer learning is used to obtain dense maps on Waymo. With the enhanced LiDAR data and the camera images, we explore early and middle fusion approaches using popular object detection models. The proposed data fusion network provides a significant improvement compared to single-modal detection at all times of the day, and outperforms previous approaches that upsample depth maps with classical image processing algorithms. Our multi-modal and multi-source approach achieves a 1.5, 7.5, and 2.1 mean AP increase at day, night, and dawn/dusk, respectively, using four different object detection meta-architectures.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据