4.6 Article

Saliency Attack: Towards Imperceptible Black-box Adversarial Attack

Publisher

ASSOC COMPUTING MACHINERY
DOI: 10.1145/3582563

Keywords

Adversarial example; black-box adversarial attack; saliency map; deep neural networks

Ask authors/readers for more resources

In this article, a method of restricting perturbations to a small salient region to generate adversarial examples that can hardly be perceived is proposed. This approach is compatible with many existing black-box attacks and significantly improves their imperceptibility. Furthermore, a new black-box attack called Saliency Attack is introduced, which aims to refine the perturbations in the salient region for better imperceptibility. Extensive experiments demonstrate that our approach achieves much better imperceptibility scores and is also robust to different detection-based defenses.
Deep neural networks are vulnerable to adversarial examples, even in the black-box setting where the attacker is only accessible to the model output. Recent studies have devised effective black-box attacks with high query efficiency. However, such performance is often accompanied by compromises in attack imperceptibility, hindering the practical use of these approaches. In this article, we propose to restrict the perturbations to a small salient region to generate adversarial examples that can hardly be perceived. This approach is readily compatible with many existing black-box attacks and can significantly improve their imperceptibility with little degradation in attack success rates. Furthermore, we propose the Saliency Attack, a newblack-box attack aiming to refine the perturbations in the salient region to achieve even better imperceptibility. Extensive experiments showthat compared to the state-of-the-art black-box attacks, our approach achievesmuch better imperceptibility scores, including most apparent distortion (MAD), L-0 and L-2 distances, and also obtains significantly better true success rate and effective query number judged by a human-like threshold on MAD. Importantly, the perturbations generated by our approach are interpretable to some extent. Finally, it is also demonstrated to be robust to different detection-based defenses.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available