4.6 Article

CSU-Net: A Context Spatial U-Net for Accurate Blood Vessel Segmentation in Fundus Images

Journal

IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS
Volume 25, Issue 4, Pages 1128-1138

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JBHI.2020.3011178

Keywords

Image segmentation; Feature extraction; Biomedical imaging; Blood vessels; Machine learning; Task analysis; Fundus images; blood vessel segmentation; CSU-Net; feature fusion; structure loss

Funding

  1. National Key Research and Development Program of China [2017YFB1302704]
  2. National Natural Science Foundation of China [61976209, 81701785]
  3. CAS International Collaboration Key Project [173211KYSB20190024]
  4. Strategic Priority Research Programof CAS [XDB32040200]

Ask authors/readers for more resources

A novel CSU-Net model is proposed for blood vessel segmentation, utilizing a two-channel encoder and feature fusion, attention skip modules to improve segmentation accuracy. Evaluation on three public datasets demonstrates that CSU-Net outperforms current state-of-the-art methods.
Blood vessel segmentation in fundus images is a critical procedure in the diagnosis of ophthalmic diseases. Recent deep learning methods achieve high accuracy in vessel segmentation but still face the challenge to segment the microvascular and detect the vessel boundary. This is due to the fact that common Convolutional Neural Networks (CNN) are unable to preserve rich spatial information and a large receptive field simultaneously. Besides, CNN models for vessel segmentation usually are trained by equal pixel level cross-entropy loss, which tend to miss fine vessel structures. In this paper, we propose a novel Context Spatial U-Net (CSU-Net) for blood vessel segmentation. Compared with the other U-Net based models, we design a two-channel encoder: a context channel with multi-scale convolution to capture more receptive field and a spatial channel with large kernel to retain spatial information. Also, to combine and strengthen the features extracted from two paths, we introduce a feature fusion module (FFM) and an attention skip module (ASM). Furthermore, we propose a structure loss, which adds a spatial weight to cross-entropy loss and guide the network to focus more on the thin vessels and boundaries. We evaluated this model on three public datasets: DRIVE, CHASE-DB1 and STARE. The results show that the CSU-Net achieves higher segmentation accuracy than the current state-of-the-art methods.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available