4.7 Article

Noniterative Deep Learning: Incorporating Restricted Boltzmann Machine Into Multilayer Random Weight Neural Networks

Journal

IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS
Volume 49, Issue 7, Pages 1299-1308

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TSMC.2017.2701419

Keywords

Deep learning (DL); generalized inverse of matrix; random weight neural network (RWNN); supervised learning; training without iteration

Funding

  1. National Natural Science Foundation of China [71371063, 61170040, 61402460, 61472257]
  2. Natural Science Foundation of SZU [2017060]
  3. Basic Research Project of Knowledge Innovation Program in Shenzhen [JCYJ20150324140036825]
  4. Guangdong Provincial Science and Technology Plan Project [2013B040403005]
  5. HD Video Research and Development Platform for Intelligent Analysis and Processing in Guangdong Engineering Technology Research Centre of Colleges and Universities [GCZX-A1409]

Ask authors/readers for more resources

A general deep learning (DL) mechanism for a multiple hidden layer feed-fiwvvard neural network contains two parts, i.e., 1) an unsupervised greedy layer-wise training and 2) a supervised fine-tuning which is usually an iterative process. Although this mechanism has been demonstrated in many fields to be able to significantly improve the generalization of neural network, there is no clear evidence to show which one of the two parts plays the essential role for the generalization improvement, resulting in an argument within the DL community. Focusing on this argument, this paper proposes a new DL approach to train multilayer feed-forward neural networks. This approach uses restricted Boltzmann machine (RBM) as the layer-wise training and uses the generalized inverse of a matrix as the supervised fine-tuning. Different from the general deep training mechanism like back-propagation (BP), the proposed approach does not need to iteratively tune the weights, and therefore, has many advantages such as quick training, better generalization, and high understandability, etc. Experimentally, the proposed approach demonstrates an excellent performance in comparison with BP-based DL and the traditional training method for multilayer random weight neural networks. To a great extent, this paper demonstrates that the supervised part plays a more important role than the unsupervised part in DL, which provides some new viewpoints for exploring the essence of DL.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available