4.6 Review

A group of industry, academic, and government experts convene in Philadelphia to explore the roots of algorithmic bias

期刊

COMMUNICATIONS OF THE ACM
卷 63, 期 5, 页码 82-89

出版社

ASSOC COMPUTING MACHINERY
DOI: 10.1145/3376898

关键词

-

资金

  1. National Science Foundation [1136993]
  2. Direct For Computer & Info Scie & Enginr
  3. Division of Computing and Communication Foundations [1136993] Funding Source: National Science Foundation

向作者/读者索取更多资源

THE LAST DECADE has seen a vast increase both in the diversity of applications to which machine learning is applied, and to the import of those applications. Machine learning is no longer just the engine behind ad placements and spam filters; it is now used to filter loan applicants, deploy police officers, and inform bail and parole decisions, among other things. The result has been a major concern for the potential for data-driven methods to introduce and perpetuate discriminatory practices, and to otherwise be unfair. And this concern has not been without reason: a steady stream of empirical findings has shown that data-driven methods can unintentionally both encode existing human biases and introduce new ones.(7,9,11,60)

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据