4.8 Article

Automatic Defect Segmentation in X-Ray Images Based on Deep Learning

期刊

IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS
卷 68, 期 12, 页码 12912-12920

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIE.2020.3047060

关键词

Image segmentation; X-ray imaging; Feature extraction; Semantics; Deep learning; Object segmentation; Manufacturing; Casting parts; computer vision; deep learning; defect segmentation; nondestructive testing (NDT); X-ray image

资金

  1. National Nature Science Foundation of China [51975518]
  2. Science Fund for Creative Research Groups of National Natural Science Foundation of China [51821093]
  3. Key Research and Development Plan of Zhejiang Province [2018C01073]
  4. Ningbo Science and Technology Plan [2019B10072]
  5. Fundamental Research Funds for the Central Universities [2019QNA4004]

向作者/读者索取更多资源

In this article, three methods were proposed to improve the segmentation of X-ray images, including enhancing contrast using CLAHE, establishing a two-stream CNN for image processing, and introducing a weighted IOU loss function. Experimental results show that these methods outperform the baseline in terms of different metrics, indicating better performance and effectiveness in object segmentation.
X-ray imaging has been broadly adopted as a nondestructive testing method for product quality inspection. Deep learning has demonstrated powerful image scene understanding capabilities. In this article, U-Net with resnet101 is taken as the baseline for defect segmentation. First, there exist gray inhomogeneous and low-contrast regions in X-ray images, which can hardly be segmented. Contrast-limited adaptive histogram equalization (CLAHE) could be used to improve the contrast and consistency of the X-ray image. A two-stream convolutional neural network (CNN) is proposed that takes the original image and CLAHE processed image as inputs to address this issue. And then in CNN, low-level feature maps are lacking semantic information, which may lead to worse results. A gated multilayer fusion module is proposed to adaptively fuse the high-level features into low-level features. Furthermore, loss functions (such as cross entropy) in semantic segmentation are usually pixel level, ignoring the regional information. A weighted intersection over union (IOU) loss function is proposed to introduce IOU information to guide the model to focus on the objects that are easy to mine. The experimental results prove that the three proposed methods have better performance than the baseline for our dataset, achieving 42.2 in mIoU, 59.2 in Dice, and 54.5%, 74.9%, and 86.3% in small, middle, and large object recall rate, respectively.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据