4.5 Article

Benchmarking Crisis in Social Media Analytics: A Solution for the Data-Sharing Problem

Journal

SOCIAL SCIENCE COMPUTER REVIEW
Volume 40, Issue 6, Pages 1496-1522

Publisher

SAGE PUBLICATIONS INC
DOI: 10.1177/08944393211012268

Keywords

social media analytics; benchmarking; social computing; reproducibility

Funding

  1. German Federal Ministry of Education and Research [FKZ 16KIS0495K]
  2. Ministry of Culture and Science of the German State of North Rhine-Westphalia [FKZ 005-1709-0001, EFRE-0801431, FKZ 005-1709-0006]
  3. European Research Center for Information Systems (ERCIS)
  4. DAAD PPP Germany-Australia [57511656]
  5. EU H2020 Program under the scheme INFRAIA-01-2018-2019: Research and Innovation action grant [871042]

Ask authors/readers for more resources

Computational social science evaluates social interaction using computational and statistical methods, with public availability of data sets being essential for reliable research. Restrictions on data sharing for social media analytics research create challenges for replicability, prompting the proposal of a new evaluation framework to address these issues.
Computational social science uses computational and statistical methods in order to evaluate social interaction. The public availability of data sets is thus a necessary precondition for reliable and replicable research. These data allow researchers to benchmark the computational methods they develop, test the generalizability of their findings, and build confidence in their results. When social media data are concerned, data sharing is often restricted for legal or privacy reasons, which makes the comparison of methods and the replicability of research results infeasible. Social media analytics research, consequently, faces an integrity crisis. How is it possible to create trust in computational or statistical analyses, when they cannot be validated by third parties? In this work, we explore this well-known, yet little discussed, problem for social media analytics. We investigate how this problem can be solved by looking at related computational research areas. Moreover, we propose and implement a prototype to address the problem in the form of a new evaluation framework that enables the comparison of algorithms without the need to exchange data directly, while maintaining flexibility for the algorithm design.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available