4.5 Article

Exploiting fusion architectures for multispectral pedestrian detection and segmentation

期刊

APPLIED OPTICS
卷 57, 期 18, 页码 D108-D116

出版社

Optica Publishing Group
DOI: 10.1364/AO.57.00D108

关键词

-

类别

资金

  1. National Natural Science Foundation of China (NSFC) [51575486, 51605428, U1664264]
  2. Fundamental Research Funds for the Central Universities

向作者/读者索取更多资源

Recent research has demonstrated that the fusion of complementary information captured by multi-modal sensors (visible and infrared cameras) enables robust pedestrian detection under various surveillance situations (e.g., daytime and nighttime). In this paper, we investigate a number of fusion architectures in an attempt to identify the optimal way of incorporating multispectral information for joint semantic segmentation and pedestrian detection. We made two important findings: (1) the sum fusion strategy, which computes the sum of two feature maps at the same spatial locations, delivers the best performance of multispectral detection, while the most commonly used concatenation fusion surprisingly performs the worst; and (2) two-stream semantic segmentation without multispectral fusion is the most effective scheme to infuse semantic information as supervision for learning human-related features. Based on these studies, we present a unified multispectral fusion framework for joint training of semantic segmentation and target detection that outperforms state-of-the-art multispectral pedestrian detectors by a large margin on the KAIST benchmark dataset. (C) 2018 Optical Society of America

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据