4.6 Article

Broad Transfer Learning Network based Li-ion battery lifetime prediction model

Journal

ENERGY REPORTS
Volume 10, Issue -, Pages 881-893

Publisher

ELSEVIER
DOI: 10.1016/j.egyr.2023.07.025

Keywords

Battery; Remaining useful life; Broad learning; Transfer learning; Feature mapping; Data augmentation

Categories

Ask authors/readers for more resources

The Broad Transfer Learning Network (BTLN) model achieves similar prediction performance as a Multi-Layer Perceptron (MLP) model using only one-third of the parameters. It combines broad learning and transfer learning techniques, and improves performance by enhancing feature extraction and increasing training efficiency. The BTLN model shows a significant improvement of 18.5% in performance compared to common neural network models.
The Broad Transfer Learning Network (BTLN) model uses only one-third of the parameters as the common neural network does to achieve the similar prediction performance that a Multi-Layer Perceptron (MLP) model has. This brings considerable benefits to AI-related studies that highly rely on computational power. This study also proposed a feature mapping technique that applied a linear transformation to the original features to improve the performance of various learning models. This study combines both broad learning and transfer learning techniques. Data augmentation is used to expand the training dataset. It proves that, under certain conditions, the model with a broader network can perform better. The broad network structure can act as an effective feature extractor. The transform learning algorithm can increase the training efficiency due to decreased trainable parameters. The performance improvement of neural network models is particularly remarkable. The performance improves by 3.5% without any change to the model architecture. In this paper, the performance of the BTLN model proposed based on Root Mean Square Error (RMSE) improves by 18.5%. Compared to common neural network models, the training parameters and overall parameters are lowered to 14.57% and 36.28%, respectively. & COPY; 2023 The Authors. Published by Elsevier Ltd. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available