4.5 Article

Label flipping attacks against Naive Bayes on spam filtering systems

Journal

APPLIED INTELLIGENCE
Volume 51, Issue 7, Pages 4503-4514

Publisher

SPRINGER
DOI: 10.1007/s10489-020-02086-4

Keywords

Spam classification; Label flipping attacks; Naive Bayes classifier; Performance evaluation

Funding

  1. Integration of Cloud Computing and Big Integration of Cloud Computing and Big Data, Innovation of Science and Education [2017A11017]
  2. Key Research, Development, and Dissemination Program of Henan Province (Science and Technology for the People) [182207310002]
  3. Key Science and Technology Project of Xinjiang Production and Construction Corps [2018AB017]

Ask authors/readers for more resources

Label flipping attack is a poisoning attack that reduces the classification performance of a model by flipping the labels of training samples. Naive Bayes algorithm demonstrates good robustness in handling issues like document classification and spam filtering. The proposed label flipping attacks effectively reduce the accuracy of various classification models.
Label flipping attack is a poisoning attack that flips the labels of training samples to reduce the classification performance of the model. Robustness is used to measure the applicability of machine learning algorithms to adversarial attack. Naive Bayes (NB) algorithm is a anti-noise and robust machine learning technique. It shows good robustness when dealing with issues such as document classification and spam filtering. Here we propose two novel label flipping attacks to evaluate the robustness of NB under label noise. For the three datasets of Spambase, TREC 2006c and TREC 2007 in the spam classification domain, our attack goal is to increase the false negative rate of NB under the influence of label noise without affecting normal mail classification. Our evaluation shows that at a noise level of 20%, the false negative rate of Spambase and TREC 2006c has increased by about 20%, and the test error of the TREC 2007 dataset has increased to nearly 30%. We compared the classification accuracy of five classic machine learning algorithms (random forest(RF), support vector machine(SVM), decision tree(DT), logistic regression(LR), and NB) and two deep learning models(AlexNet, LeNet) under the proposed label flipping attacks. The experimental results show that two label noises are suitable for various classification models and effectively reduce the accuracy of the models.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available