4.5 Article

Predictive Coding Approximates Backprop Along Arbitrary Computation Graphs

Journal

NEURAL COMPUTATION
Volume 34, Issue 6, Pages 1329-1368

Publisher

MIT PRESS
DOI: 10.1162/neco_a_01497

Keywords

-

Funding

  1. EPSRC
  2. Dr. Mortimer and Theresa Sackler Foundation
  3. School of Engineering and Informatics at the University of Sussex
  4. BBRSC [BB/P022197/1]
  5. National Institutes of Natural Sciences, Japan [01112005]

Ask authors/readers for more resources

Backpropagation of error can be approximated using predictive coding, a biologically plausible process theory of cortical computation. This result allows for the translation of core machine learning architectures into predictive coding equivalents, which perform equivalently to backprop on challenging machine learning benchmarks using only local and Hebbian plasticity.
Backpropagation of error (backprop) is a powerful algorithm for training machine learning architectures through end-to-end differentiation. Recently it has been shown that backprop in multilayer perceptrons (MLPs) can be approximated using predictive coding, a biologically plausible process theory of cortical computation that relies solely on local and Hebbian updates. The power of backprop, however, lies not in its instantiation in MLPs but in the concept of automatic differentiation, which allows for the optimization of any differentiable program expressed as a computation graph. Here, we demonstrate that predictive coding converges asymptotically (and in practice, rapidly) to exact backprop gradients on arbitrary computation graphs using only local learning rules. We apply this result to develop a straightforward strategy to translate core machine learning architectures into their predictive coding equivalents. We construct predictive coding convolutional neural networks, recurrent neural networks, and the more complex long short-term memory, which include a nonlayer-like branching internal graph structure and multiplicative interactions. Our models perform equivalently to backprop on challenging machine learning benchmarks while using only local and (mostly) Hebbian plasticity. Our method raises the potential that standard machine learning algorithms could in principle be directly implemented in neural circuitry and may also contribute to the development of completely distributed neuromorphic architectures.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available