4.7 Article

Balancing Learning Model Privacy, Fairness, and Accuracy With Early Stopping Criteria

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2021.3129592

Keywords

Training; Privacy; Deep learning; Costs; Analytical models; Stability criteria; Stochastic processes; Deep learning; differential privacy (DP); early stopping criteria; machine learning fairness; stochastic gradient descent

Ask authors/readers for more resources

As deep learning models mature, finding the ideal tradeoff between accuracy, fairness, and privacy is critical. Privacy and fairness can affect the accuracy of models, so balancing these needs is important. By implementing differentially private stochastic gradient descent (DP-SGD) in deep neural network models, privacy and fairness can be indirectly managed. The number of training epochs plays a central role in striking a balance between accuracy, fairness, and privacy. Based on this observation, two early stopping criteria are designed to help analysts achieve their ideal tradeoff.
As deep learning models mature, one of the most prescient questions we face is: what is the ideal tradeoff between accuracy, fairness, and privacy (AFP)? Unfortunately, both the privacy and the fairness of a model come at the cost of its accuracy. Hence, an efficient and effective means of fine-tuning the balance between this trinity of needs is critical. Motivated by some curious observations in privacy-accuracy tradeoffs with differentially private stochastic gradient descent (DP-SGD), where fair models sometimes result, we conjecture that fairness might be better managed as an indirect byproduct of this process. Hence, we conduct a series of analyses, both theoretical and empirical, on the impacts of implementing DP-SGD in deep neural network models through gradient clipping and noise addition. The results show that, in deep learning, the number of training epochs is central to striking a balance between AFP because DP-SGD makes the training less stable, providing the possibility of model updates at a low discrimination level without much loss in accuracy. Based on this observation, we designed two different early stopping criteria to help analysts choose the optimal epoch at which to stop training a model so as to achieve their ideal tradeoff. Extensive experiments show that our methods can achieve an ideal balance between AFP.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available