4.5 Article

AIPA: An Adversarial Imperceptible Patch Attack on Medical Datasets and its Interpretability

Journal

COMPUTERS & SECURITY
Volume 136, Issue -, Pages -

Publisher

ELSEVIER ADVANCED TECHNOLOGY
DOI: 10.1016/j.cose.2023.103574

Keywords

Deep learning; Adversarial attack; Interpretability; Patch attack

Ask authors/readers for more resources

This paper proposes a novel adversarial attack technique that can synthesize adversarial images to mislead deep learning models, and also studies interpretability plots. The research findings show that the proposed attack technique influences the interpretability plots, regardless of the success of the attack.
Deep learning is one of the most prominent computational techniques used for automated disease detection in the medical domain. In the field of deep learning, the performance and reliability of deep learning models have been compromised due to adversarial attacks. In this work, a novel Adversarial Imperceptible Patch Attack (AIPA) is proposed. Adversarial noise, which is created as a small rectangular patch of noise, is added to an original image to synthesise the adversarial image. The Diabetic Retinopathy 2015 Data Colored Resized dataset and the SARS-COV-2 CT-Scan dataset have been used in this experimentation. It is found that for both the datasets, the adversarial image synthesised by the proposed technique is capable of misleading a customised VGG16 in terms of its classification. Interpretability plots generated using the Gradient Shap, Integrated Gradients, Occlusion and Saliency techniques are also studied. The proposed adversarial attack has influenced the interpretability plots, irrespective of whether the adversarial attack is successful in misclassification or not. When the attack is successful with respect to classification, the interpretability plots seem to favour misprediction. In addition, when the attack is unsuccessful, interpretability plots are inconsistent. The susceptibility of the deep learning models revealed by this work would be beneficial for the research community to devise better defence mechanisms and interpretability methods.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available