4.6 Review Book Chapter

Computational Principles of Supervised Learning in the Cerebellum

Journal

ANNUAL REVIEW OF NEUROSCIENCE, VOL 41
Volume 41, Issue -, Pages 233-253

Publisher

ANNUAL REVIEWS
DOI: 10.1146/annurev-neuro-080317-061948

Keywords

machine learning; decorrelation; consolidation; plasticity; climbing fiber; Purkinje cell

Categories

Funding

  1. NATIONAL INSTITUTE OF MENTAL HEALTH [R01MH093727, RF1MH114269] Funding Source: NIH RePORTER
  2. NATIONAL INSTITUTE ON ALCOHOL ABUSE AND ALCOHOLISM [R21AA025572] Funding Source: NIH RePORTER

Ask authors/readers for more resources

Supervised learning plays a key role in the operation of many biological and artificial neural networks. Analysis of the computations underlying supervised learning is facilitated by the relatively simple and uniform architecture of the cerebellum, a brain area that supports numerous motor, sensory, and cognitive functions. We highlight recent discoveries indicating that the cerebellum implements supervised learning using the following organizational principles: (a) extensive preprocessing of input representations (i.e., feature engineering), (b) massively recurrent circuit architecture, (c) linear input-output computations, (d) sophisticated instructive signals that can be regulated and are predictive, (e) adaptive mechanisms of plasticity with multiple timescales, and (f) task-specific hardware specializations. The principles emerging from studies of the cerebellum have striking parallels with those in other brain areas and in artificial neural networks, as well as some notable differences, which can inform future research on supervised learning and inspire next-generation machine-based algorithms.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available