4.7 Article

Vision transformer-based autonomous crack detection on asphalt and concrete surfaces

期刊

AUTOMATION IN CONSTRUCTION
卷 140, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.autcon.2022.104316

关键词

Crack detection; Deep learning; Vision transformer; Convolutional neural network; Human recognition system

资金

  1. School of Civil Engineering at the University of Sydney
  2. University of Sydney

向作者/读者索取更多资源

This research proposes a ViT-based framework for crack detection on asphalt and concrete surfaces, achieving enhanced real-world crack segmentation performance through transfer learning and IoU loss function. Compared to CNN-based models, TransUNet with a CNN-ViT backbone shows better average IoU on small and multiscale crack semantics and ViT helps the encoder-decoder network exhibit robust performance against various noisy signals.
Previous research has shown the high accuracy of convolutional neural networks (CNNs) in asphalt and concrete crack detection in controlled conditions. Yet, human-like generalisation remains a significant challenge for industrial applications where the range of conditions varies significantly. Given the intrinsic biases of CNNs, this paper proposes a vision transformer (ViT)-based framework for crack detection on asphalt and concrete surfaces. With transfer learning and the differentiable intersection over union (IoU) loss function, the encoder-decoder network equipped with ViT could achieve an enhanced real-world crack segmentation performance. Compared to the CNN-based models (DeepLabv3+ and U-Net), TransUNet with a CNN-ViT backbone achieved up to -61% and -3.8% better mean IoU on the original images of the respective datasets with very small and multiscale crack semantics. Moreover, ViT assisted the encoder-decoder network to show a robust performance against various noisy signals where the mean Dice score attained by the CNN-based models significantly dropped (<10%).

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据