3.8 Proceedings Paper

Image classification based on RESNET

Publisher

IOP PUBLISHING LTD
DOI: 10.1088/1742-6596/1634/1/012110

Keywords

-

Ask authors/readers for more resources

at present, neural networks are becoming more and more complex, from several layers to dozens of layers or even more than 100 layers.The main advantage of deep network is that it can express very complex functions. It can learn features from different levels of abstraction, such as edge features at lower levels and complex features at higher levels.However, the use of deep networks is not always effective, because there is a very big obstacle - the disappearance of gradients: in very deep networks, gradient signals tend to approach zero very quickly, which makes the gradient descent process extremely slow. Specifically, in the process of gradient descent, the weight matrix product operation must be carried out in every step of back propagation from the last layer to the first layer, so that the gradient will drop exponentially to 0.(in rare cases, there is the problem of gradient explosion, that is, the gradient grows exponentially to the overflow in the process of propagation). Therefore, in the process of training, it will be found that with the increase of the number of layers, the rate of gradient decrease increases.Therefore, by deepening the network, although it can express any complex function, but in fact, with the increase of network layers, we are more and more difficult to train the network, until the proposal of residual network, which makes it possible to train deeper network[1].

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available