4.7 Article

Detection defense against adversarial attacks with saliency map

Journal

INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS
Volume 37, Issue 12, Pages 10193-10210

Publisher

WILEY-HINDAWI
DOI: 10.1002/int.22458

Keywords

adversarial defense; adversarial example; deep neural network; machine learning; saliency map

Funding

  1. National Natural Science Foundation of China NSFC [62072343, U1636219]
  2. National Key Research and Development Program of China [2019QY (Y)0206]

Ask authors/readers for more resources

This paper discusses the saliency map method for enhancing model interpretability, as well as a novel approach combined with additional noises and inconsistency strategy to detect adversarial examples. Experimental results demonstrate that the proposed method effectively detects adversarial attacks with high success rate across common datasets and models, showing its generality compared to existing state-of-the-art techniques.
It is well established that neural networks are vulnerable to adversarial examples, which are almost imperceptible on human vision and can cause the deep models misbehave. Such phenomenon may lead to severely inestimable consequences in the safety and security critical applications. Existing defenses are trend to harden the robustness of models against adversarial attacks, for example, adversarial training technology. However, these are usually intractable to implement due to the high cost of retraining and the cumbersome operations of altering the model architecture or parameters. In this paper, we discuss the saliency map method from the view of enhancing model interpretability, it is similar to introducing the mechanism of the attention to the model, so as to comprehend the progress of object identification by the deep networks. We then propose a novel method combined with additional noises and utilize the inconsistency strategy to detect adversarial examples. Our experimental results of some representative adversarial attacks on common data sets including ImageNet and popular models show that our method can detect all the attacks with high detection success rate effectively. We compare it with the existing state-of-the-art technique, and the experiments indicate that our method is more general.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available