4.7 Article

Continual Learning Using Bayesian Neural Networks

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2020.3017292

Keywords

Task analysis; Adaptation models; Training; Bayes methods; Modeling; Uncertainty; Gaussian distribution; Bayesian neural networks (BNNs); catastrophic forgetting; continual learning; incremental learning; uncertainty

Funding

  1. Care Research and Technology Centre at the U.K. Dementia Research Institute (U.K. DRI)
  2. MRC [UKDRI-7002] Funding Source: UKRI

Ask authors/readers for more resources

Continual learning models face the challenge of catastrophic forgetting in dynamic environments. To address this issue, a method called continual Bayesian learning networks (CBLNs) is proposed, which optimizes resource allocation and weight selection by maintaining a mixture of Gaussian posterior distributions to solve the problem of forgetting previously learned tasks.
Continual learning models allow them to learn and adapt to new changes and tasks over time. However, in continual and sequential learning scenarios, in which the models are trained using different data with various distributions, neural networks (NNs) tend to forget the previously learned knowledge. This phenomenon is often referred to as catastrophic forgetting. The catastrophic forgetting is an inevitable problem in continual learning models for dynamic environments. To address this issue, we propose a method, called continual Bayesian learning networks (CBLNs), which enables the networks to allocate additional resources to adapt to new tasks without forgetting the previously learned tasks. Using a Bayesian NN, CBLN maintains a mixture of Gaussian posterior distributions that are associated with different tasks. The proposed method tries to optimize the number of resources that are needed to learn each task and avoids an exponential increase in the number of resources that are involved in learning multiple tasks. The proposed method does not need to access the past training data and can choose suitable weights to classify the data points during the test time automatically based on an uncertainty criterion. We have evaluated the method on the MNIST and UCR time-series data sets. The evaluation results show that the method can address the catastrophic forgetting problem at a promising rate compared to the state-of-the-art models.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available