4.7 Article

Perturbation Inactivation Based Adversarial Defense for Face Recognition

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIFS.2022.3195384

Keywords

Face recognition; Perturbation methods; Robustness; Immune system; Principal component analysis; Deep learning; Training; Adversarial machine learning; deep learning; graph neural network; face recognition

Funding

  1. National Natural Science Foundation of China [U1836217, 62006225, 61622310, 62071468]
  2. Strategic Priority Research Program of Chinese Academy of Sciences [XDA27040700]

Ask authors/readers for more resources

This paper proposes a plug-and-play defense method called perturbation inactivation (PIN) to improve the robustness of deep learning-based face recognition models against adversarial attacks. The method estimates an immune space in which adversarial perturbations have less impact on the recognition model and restricts the perturbations to this subspace. The method outperforms state-of-the-art defense methods and demonstrates good generalization capabilities in exhaustive experiments. It can also be easily applied to existing commercial APIs and face recognition systems.
Deep learning-based face recognition models are vulnerable to adversarial attacks. To curb these attacks, most defense methods aim to improve the robustness of recognition models against adversarial perturbations. However, the generalization capacities of these methods are quite limited. In practice, they are still vulnerable to unseen adversarial attacks. Deep learning models are fairly robust to general perturbations, such as Gaussian noises. A straightforward approach is to inactivate the adversarial perturbations so that they can be easily handled as general perturbations. In this paper, a plug-and-play adversarial defense method, named perturbation inactivation (PIN), is proposed to inactivate adversarial perturbations for adversarial defense. We discover that the perturbations in different subspaces have different influences on the recognition model. There should be a subspace, called the immune space, in which the perturbations have fewer adverse impacts on the recognition model than in other subspaces. Hence, our method estimates the immune space and inactivates the adversarial perturbations by restricting them to this subspace. The proposed method can be generalized to unseen adversarial perturbations since it does not rely on a specific kind of adversarial attack method. This approach not only outperforms several state-of-the-art adversarial defense methods but also demonstrates a superior generalization capacity through exhaustive experiments. Moreover, the proposed method can be successfully applied to four commercial APIs without additional training, indicating that it can be easily generalized to existing face recognition systems.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available