4.7 Article

Multi-Modal Retinal Image Classification With Modality-Specific Attention Network

Journal

IEEE TRANSACTIONS ON MEDICAL IMAGING
Volume 40, Issue 6, Pages 1591-1602

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TMI.2021.3059956

Keywords

Retina; Deep learning; Feature extraction; Biomedical imaging; Optical imaging; Image segmentation; Training; Fundus photography; optical coherence tomography; classification; multi-modal; attention; convolutional neural network

Funding

  1. National Natural Science Foundation of China [61922029]
  2. Science and Technology Plan Project Fund of Hunan Province [2019RS2016]

Ask authors/readers for more resources

In this study, a novel modality-specific attention network (MSAN) was proposed for multi-modal retinal image classification, effectively utilizing diagnostic features from fundus and OCT images. The network includes two attention modules to extract features from different imaging modalities for a more accurate diagnosis.
Recently, automatic diagnostic approaches have been widely used to classify ocular diseases. Most of these approaches are based on a single imaging modality (e.g., fundus photography or optical coherence tomography (OCT)), which usually only reflect the oculopathy to a certain extent, and neglect the modality-specific information among different imaging modalities. This paper proposes a novel modality-specific attention network (MSAN) for multi-modal retinal image classification, which can effectively utilize the modality-specific diagnostic features from fundus and OCT images. The MSAN comprises two attention modules to extract the modality-specific features from fundus and OCT images, respectively. Specifically, for the fundus image, ophthalmologists need to observe local and global pathologies at multiple scales (e.g., from microaneurysms at the micrometer level, optic disc at millimeter level to blood vessels through the whole eye). Therefore, we propose a multi-scale attention module to extract both the local and global features from fundus images. Moreover, large background regions exist in the OCT image, which is meaningless for diagnosis. Thus, a region-guided attention module is proposed to encode the retinal layer-related features and ignore the background in OCT images. Finally, we fuse the modality-specific features to form a multi-modal feature and train the multi-modal retinal image classification network. The fusion of modality-specific features allows the model to combine the advantages of fundus and OCT modality for a more accurate diagnosis. Experimental results on a clinically acquired multi-modal retinal image (fundus and OCT) dataset demonstrate that our MSAN outperforms other well-known single-modal and multi-modal retinal image classification methods.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available