4.7 Article

DenseSPH-YOLOv5: An automated damage detection model based on DenseNet and Swin-Transformer prediction head-enabled YOLOv5 with attention mechanism

Journal

ADVANCED ENGINEERING INFORMATICS
Volume 56, Issue -, Pages -

Publisher

ELSEVIER SCI LTD
DOI: 10.1016/j.aei.2023.102007

Keywords

Automated road damage detection; You Only Look Once (YOLOv5) algorithm; Swin transformer; object detection; Computer vision; Deep learning

Ask authors/readers for more resources

In order to address the deficiencies of existing DL-based damage detection models in complex and noisy environments, the authors propose a real-time high-performance damage detection model called DenseSPH-YOLOv5. By integrating DenseNet blocks with the backbone and implementing convolutional block attention modules (CBAM), the model achieves superior deep spatial feature extraction in challenging environments. Additional feature fusion layers and a Swin-Transformer Prediction Head (SPH) are also added to improve the efficiency of multi-scale object detection and reduce computational complexity.
Objective. Computer vision-based up-to-date accurate damage classification and localization are of decisive importance for infrastructure monitoring, safety, and the serviceability of civil infrastructure. Current state-of-the-art deep learning (DL)-based damage detection models, however, often lack superior feature extraction capability in complex and noisy environments, limiting the development of accurate and reliable object distinction. Method. To this end, we present DenseSPH-YOLOv5, a real-time DL-based high-performance damage detection model where DenseNet blocks have been integrated with the backbone to improve in preserving and reusing critical feature information. Additionally, convolutional block attention modules (CBAM) have been implemented to improve attention performance mechanisms for strong and discriminating deep spatial feature extraction that results in superior detection under various challenging environments. Moreover, an additional feature fusion layers and a Swin-Transformer Prediction Head (SPH) have been added leveraging advanced self-attention mechanism for more efficient detection of multiscale object sizes and simultaneously reducing the computational complexity.Results. Evaluating the model performance in large-scale Road Damage Dataset (RDD-2018), at a detection rate of 62.4 FPS, DenseSPH-YOLOv5 obtains a mean average precision (mAP) value of 85.25%, F1-score of 81.18%, and precision (P) value of 89.51% outperforming current state-of-the-art models.Significance. The present research provides an effective and efficient damage localization model addressing the shortcoming of existing DL-based damage detection models by providing highly accurate localized bounding box prediction. Current work constitutes a step towards an accurate and robust automated damage detection system in real-time in-field applications.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available