3.8 Proceedings Paper

An Investigation of Critical Issues in Bias Mitigation Techniques

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/WACV51458.2022.00257

Keywords

-

Funding

  1. DARPA/SRI Lifelong Learning Machines program [HR0011-18-C-0051]
  2. AFOSR grant [FA9550-18-1-0121]
  3. NSF [1909696]
  4. Direct For Computer & Info Scie & Enginr
  5. Div Of Information & Intelligent Systems [1909696] Funding Source: National Science Foundation

Ask authors/readers for more resources

A critical problem in deep learning is that systems learn inappropriate biases, resulting in their inability to perform well on minority groups. Multiple algorithms have been developed to mitigate bias but their effectiveness remains unclear.
A critical problem in deep learning is that systems learn inappropriate biases, resulting in their inability to perform well on minority groups. This has led to the creation of multiple algorithms that endeavor to mitigate bias. However, it is not clear how effective these methods are. This is because study protocols differ among papers, systems are tested on datasets that fail to test many forms of bias, and systems have access to hidden knowledge or are tuned specifically to the test set. To address this, we introduce an improved evaluation protocol, sensible metrics, and a new dataset, which enables us to ask and answer critical questions about bias mitigation algorithms. We evaluate seven state-of-the-art algorithms using the same network architecture and hyperparameter selection policy across three benchmark datasets. We introduce a new dataset called Biased MNIST that enables assessment of robustness to multiple bias sources. We use Biased MNIST and a visual question answering (VQA) benchmark to assess robustness to hidden biases. Rather than only tuning to the test set distribution, we study robustness across different tuning distributions, which is critical because for many applications the test distribution may not be known during development. We find that algorithms exploit hidden biases, are unable to scale to multiple forms of bias, and are highly sensitive to the choice of tuning set. Based on our findings, we implore the community to adopt more rigorous assessment of future bias mitigation methods. All data, code, and results are publicly available(1).

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available