4.7 Article

Mode-assisted unsupervised learning of restricted Boltzmann machines

期刊

COMMUNICATIONS PHYSICS
卷 3, 期 1, 页码 -

出版社

NATURE PUBLISHING GROUP
DOI: 10.1038/s42005-020-0373-8

关键词

-

资金

  1. DARPA [HR00111990069]
  2. DoD-SMART fellowship
  3. Center for Memory and Recording Research at the University of California, San Diego

向作者/读者索取更多资源

Restricted Boltzmann machines (RBMs) are a powerful class of generative models, but their training requires computing a gradient that, unlike supervised backpropagation on typical loss functions, is notoriously difficult even to approximate. Here, we show that properly combining standard gradient updates with an off-gradient direction, constructed from samples of the RBM ground state (mode), improves training dramatically over traditional gradient methods. This approach, which we call 'mode-assisted training', promotes faster training and stability, in addition to lower converged relative entropy (KL divergence). We demonstrate its efficacy on synthetic datasets where we can compute KL divergences exactly, as well as on a larger machine learning standard (MNIST). The proposed mode-assisted training can be applied in conjunction with any given gradient method, and is easily extended to more general energy-based neural network structures such as deep, convolutional and unrestricted Boltzmann machines. While widely adopted, contrastive divergence methods for Restricted Boltzmann Machines typically result in poor representations of the data distribution. Here, the authors propose an unsupervised training where gradient-descent is combined with the Machine's mode samples, significantly improving the final model quality.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据