4.6 Article

Deep Learning-Based RGB-D Fusion for Multimodal Condition Assessment of Civil Infrastructure

期刊

出版社

ASCE-AMER SOC CIVIL ENGINEERS
DOI: 10.1061/JCCEE5.CPENG-5197

关键词

-

向作者/读者索取更多资源

Recent advancements in computer vision and deep learning have expanded the possibilities for autonomous condition assessment of civil infrastructure based on vision. However, the existing literature shows that most vision-based inspection techniques only rely on color information, leading to a loss of distance and scale information. This study addresses the knowledge gap by incorporating depth fusion into a semantic segmentation model. The study explores different encoding techniques for depth data and investigates fusion strategies for RGB and depth data. Overall, feature-level fusion is found to be the most effective, improving deep learning-based damage segmentation algorithms by up to 25% without increasing computation time. A novel volumetric damage quantification approach is also introduced. This study is expected to advance infrastructure resilience and maintenance.
Recent advancements in the areas of computer vision and deep learning have broadened the scope of vision-based autonomous condition assessment of civil infrastructure. However, a review of available literature suggests that most of the existing vision-based inspection techniques rely only on color information, due to the immediate availability of inexpensive high-resolution cameras. Regular cameras translate a 3D scene to a 2D space, which leads to a loss of information vis-a-vis distance and scale. This imposes a barrier to the realization of the full potential of vision-based techniques. In this regard, the structural health monitoring community is yet to benefit from the new opportunities that commercially-available low-cost depth sensors offer. This study aims at filling this knowledge gap by incorporating depth fusion into an encoder-decoder-based semantic segmentation model. Advanced computer graphics approaches are exploited to generate a database of paired RGB and depth images representing various damage categories that are commonly observed in reinforced concrete (RC) buildings, namely, spalling, spalling with exposed rebars, and severely buckled rebars. A number of encoding techniques are explored for representing the depth data. Additionally, various schemes for the data-level, feature-level, and decision-level fusions of RGB and depth data are investigated to identify the best fusion strategy. Overall, it was observed that feature-level fusion is the most effective and can enhance the performance of deep learning-based damage segmentation algorithms by up to 25% without any appreciable increase in the computation time. Moreover, a novel volumetric damage quantification approach is introduced, which is robust against perspective distortion. This study is believed to advance the frontiers of infrastructure resilience and maintenance.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据