4.6 Article

The False positive problem of automatic bot detection in social science research

Journal

PLOS ONE
Volume 15, Issue 10, Pages -

Publisher

PUBLIC LIBRARY SCIENCE
DOI: 10.1371/journal.pone.0241045

Keywords

-

Funding

  1. Ministry of Science and Technology, Taiwan (R.O.C) [108-2410-H-002 -007 -MY2]

Ask authors/readers for more resources

The identification of bots is an important and complicated task. The bot classifierBotometerwas successfully introduced as a way to estimate the number of bots in a given list of accounts and, as a consequence, has been frequently used in academic publications. Given its relevance for academic research and our understanding of the presence of automated accounts in any given Twitter discourse, we are interested inBotometer's diagnostic ability over time. To do so, we collected theBotometerscores for five datasets (three verified as bots, two verified as human; n = 4,134) in two languages (English/German) over three months. We show that theBotometerscores are imprecise when it comes to estimating bots; especially in a different language. We further show in an analysis ofBotometerscores over time thatBotometer's thresholds, even when used very conservatively, are prone to variance, which, in turn, will lead to false negatives (i.e., bots being classified as humans) and false positives (i.e., humans being classified as bots). This has immediate consequences for academic research as most studies in social science using the tool will unknowingly count a high number of human users as bots and vice versa. We conclude our study with a discussion about how computational social scientists should evaluate machine learning systems that are developed for identifying bots.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available