4.6 Article

Batch-Normalized Deep Recurrent Neural Network for High-Speed Nonlinear Circuit Macromodeling

Journal

IEEE TRANSACTIONS ON MICROWAVE THEORY AND TECHNIQUES
Volume 70, Issue 11, Pages 4857-4868

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TMTT.2022.3200071

Keywords

Integrated circuit modeling; Training; Solid modeling; Neural networks; Behavioral sciences; Nonlinear circuits; Recurrent neural networks; Batch normalization (BN); computer-aided design (CAD); modeling; nonlinear circuits; recurrent neural network (RNN)

Funding

  1. National Natural Science Foundation of China [62101382, 61901010]
  2. Scientific Research Project of Beijing Educational Committee [KM202110005029]

Ask authors/readers for more resources

The article introduces the application of batch normalization to deep RNN for faster training and improved model accuracy. By modifying the distribution of internal nodes, training speed can be increased, internal covariance shift can be reduced, and the training of deep neural networks can be accelerated.
In order to model high-speed nonlinear circuits, recurrent neural network (RNN) has been widely used in computer-aided design (CAD) area to achieve high performance and fast models compared with the existing models. Despite their advantages, they still have challenges such as large training time and limited test accuracy. In this article, the batch normalization (BN) method is applied to deep RNN leading to a much shorter training time and more accurate models compared with the conventional RNN. The proposed BN-RNN method works by modifying the distribution of the internal nodes of a deep network in the training course as an internal auxiliary shift yielding a much faster training. Indeed, the internal covariance shift will be reduced and the training of deep neural networks will be accelerated via a normalization step applied to the layers of RNN. BN-RNN, moreover, has a beneficial effect on gradient flow through the grid by reducing the dependence of gradients on the scale of network parameters or their initial values. This provides a much better learning process without the risk of divergence. For verifying the proposed method, time-domain modeling of three high-speed nonlinear circuits operating at the GHz region is provided. Comparisons of the training and test errors between RNN and BN-RNN, and evaluation time comparisons between transistor level and the BN-RNN-based models for these circuits prove the higher speed of the models obtained from the BN-RNN method. In addition, it is shown that training using the proposed method requires much less CPU time and number of epochs.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available