4.7 Article

f-AnoGAN: Fast unsupervised anomaly detection with generative adversarial networks

期刊

MEDICAL IMAGE ANALYSIS
卷 54, 期 -, 页码 30-44

出版社

ELSEVIER
DOI: 10.1016/j.media.2019.01.010

关键词

Anomaly detection; Wasserstein generative adversarial network; Unsupervised learning; Optical coherence tomography

资金

  1. Austrian Federal Ministry of Science, Research and Economy (CDL OPTIMA)
  2. IBM
  3. Austrian Science Fund FWF [12714-B31]
  4. Austrian National Bank Anniversary Fund OeNB [15356, 15929]
  5. NVIDIA Corporation
  6. Austrian Science Fund (FWF) [I2714] Funding Source: Austrian Science Fund (FWF)

向作者/读者索取更多资源

Obtaining expert labels in clinical imaging is difficult since exhaustive annotation is time-consuming. Furthermore, not all possibly relevant markers may be known and sufficiently well described a priori to even guide annotation. While supervised learning yields good results if expert labeled training data is available, the visual variability, and thus the vocabulary of findings, we can detect and exploit, is limited to the annotated lesions. Here, we present fast AnoGAN (f-AnoGAN), a generative adversarial network (GAN) based unsupervised learning approach capable of identifying anomalous images and image segments, that can serve as imaging biomarker candidates. We build a generative model of healthy training data, and propose and evaluate a fast mapping technique of new data to the GAN's latent space. The mapping is based on a trained encoder, and anomalies are detected via a combined anomaly score based on the building blocks of the trained model - comprising a discriminator feature residual error and an image reconstruction error. In the experiments on optical coherence tomography data, we compare the proposed method with alternative approaches, and provide comprehensive empirical evidence that f-AnoGAN outperforms alternative approaches and yields high anomaly detection accuracy. In addition, a visual Turing test with two retina experts showed that the generated images are indistinguishable from real normal retinal OCT images. The f-AnoGAN code is available at https://github.com/tSchlegl/f-AnoGAN. (C) 2019 Published by Elsevier B.V.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据