4.6 Review

Adversarial example detection for DNN models: a review and experimental comparison

期刊

ARTIFICIAL INTELLIGENCE REVIEW
卷 55, 期 6, 页码 4403-4462

出版社

SPRINGER
DOI: 10.1007/s10462-021-10125-w

关键词

Adversarial examples; Adversarial attacks; Detection; Deep learning; Security

资金

  1. Region Bretagne (Brittany region), France
  2. direction generale de l'armement (DGA)

向作者/读者索取更多资源

Deep learning has achieved great success in many human-related tasks and has been widely adopted in computer vision applications. Defense and detection of adversarial examples pose challenges in safety-critical applications. Existing literature proposes various countermeasures, but there is limited focus on detection methods for adversarial examples. This paper provides a survey on test-time evasion attack detection methods for neural network classifiers in image classification tasks, including experimental results of eight state-of-the-art detectors on four datasets.
Deep learning (DL) has shown great success in many human-related tasks, which has led to its adoption in many computer vision based applications, such as security surveillance systems, autonomous vehicles and healthcare. Such safety-critical applications have to draw their path to success deployment once they have the capability to overcome safety-critical challenges. Among these challenges are the defense against or/and the detection of the adversarial examples (AEs). Adversaries can carefully craft small, often imperceptible, noise called perturbations to be added to the clean image to generate the AE. The aim of AE is to fool the DL model which makes it a potential risk for DL applications. Many test-time evasion attacks and countermeasures, i.e., defense or detection methods, are proposed in the literature. Moreover, few reviews and surveys were published and theoretically showed the taxonomy of the threats and the countermeasure methods with little focus in AE detection methods. In this paper, we focus on image classification task and attempt to provide a survey for detection methods of test-time evasion attacks on neural network classifiers. A detailed discussion for such methods is provided with experimental results for eight state-of-the-art detectors under different scenarios on four datasets. We also provide potential challenges and future perspectives for this research direction.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据