3.8 Proceedings Paper

Gradient-Free Neural Network Training Based on Deep Dictionary Learning with the Log Regularizer

Journal

PATTERN RECOGNITION AND COMPUTER VISION, PT IV
Volume 13022, Issue -, Pages 561-574

Publisher

SPRINGER INTERNATIONAL PUBLISHING AG
DOI: 10.1007/978-3-030-88013-2_46

Keywords

Deep dictionary learning; log regularizer; Block coordinate descent; Sparse proximal operator; Gradient-free network

Ask authors/readers for more resources

This paper introduces a gradient-free neural network training method by using deep dictionary learning and logarithm function as sparse regularizer for feature extraction in network training. Proximal block coordinate descent method and log-thresholding operator are employed for optimizing non-convex and nonsmooth subproblems.
Gradient-free neural network training is attracting increasing attentions, which efficiently to avoid the gradient vanishing issue in traditional neural network training with gradient-based methods. The state-of-the-art gradient-free methods introduce a quadratic penalty or use an equivalent approximation of the activation function to achieve the training process without gradients, but they are hardly to mine effective signal features since the activation function is a limited nonlinear transformation. In this paper, we first propose to construct the neural network training as a deep dictionary learning model for achieving the gradientfree training of the network. To further enhance the ability of feature extraction in network training based on gradient-free method, we introduce the logarithm function as a sparsity regularizer which introduces accurate sparse activations on the hidden layer except for the last layer. Then, we employ a proximal block coordinate descent method to forward update the variables of each layer and apply the log-thresholding operator to achieve the optimization of the non-convex and non-smooth subproblems. Finally, numerical experiments conducted on several publicly available datasets prove the sparse representation of inputs is effective for gradient-free neural network training.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available