4.7 Article

On the Performance of One-Stage and Two-Stage Object Detectors in Autonomous Vehicles Using Camera Data

期刊

REMOTE SENSING
卷 13, 期 1, 页码 -

出版社

MDPI
DOI: 10.3390/rs13010089

关键词

autonomous vehicles; convolutional neural networks; deep learning; object detection; transfer learning

资金

  1. Spanish Ministry of Economy and Competitiveness [TIN2017-88209-C2-2-R]
  2. Andalusian Regional Government [US-1263341, P18-RT-2778]

向作者/读者索取更多资源

In this study, the performance of existing 2D detection systems for self-driving vehicles on a multi-class problem was evaluated and compared in different scenarios. Despite the increasing popularity of one-stage detectors, it was found that two-stage detectors still provide the most robust performance.
Object detection using remote sensing data is a key task of the perception systems of self-driving vehicles. While many generic deep learning architectures have been proposed for this problem, there is little guidance on their suitability when using them in a particular scenario such as autonomous driving. In this work, we aim to assess the performance of existing 2D detection systems on a multi-class problem (vehicles, pedestrians, and cyclists) with images obtained from the on-board camera sensors of a car. We evaluate several one-stage (RetinaNet, FCOS, and YOLOv3) and two-stage (Faster R-CNN) deep learning meta-architectures under different image resolutions and feature extractors (ResNet, ResNeXt, Res2Net, DarkNet, and MobileNet). These models are trained using transfer learning and compared in terms of both precision and efficiency, with special attention to the real-time requirements of this context. For the experimental study, we use the Waymo Open Dataset, which is the largest existing benchmark. Despite the rising popularity of one-stage detectors, our findings show that two-stage detectors still provide the most robust performance. Faster R-CNN models outperform one-stage detectors in accuracy, being also more reliable in the detection of minority classes. Faster R-CNN Res2Net-101 achieves the best speed/accuracy tradeoff but needs lower resolution images to reach real-time speed. Furthermore, the anchor-free FCOS detector is a slightly faster alternative to RetinaNet, with similar precision and lower memory usage.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据