4.7 Article

A New Macromodeling Method Based on Deep Gated Recurrent Unit Regularized With Gaussian Dropout for Nonlinear Circuits

Journal

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TCSI.2023.3264616

Keywords

Integrated circuit modeling; Solid modeling; Recurrent neural networks; Logic gates; Neural networks; Nonlinear circuits; Training; Computer-aided design (CAD); deep neural network; gated recurrent unit (GRU); Gaussian dropout; macromodeling; nonlinear circuits; recurrent neural network (RNN)

Ask authors/readers for more resources

This paper introduces the use of deep gated recurrent unit (Deep GRU) as a novel macromodeling approach for nonlinear circuits. The GRU, similar to LSTM, has gating units that control information flow and address the vanishing gradient problem. The proposed method outperforms the conventional LSTM macromodeling method in terms of accuracy and parameter efficiency, and the application of Gaussian dropout on deep GRU further improves its performance and reduces overfitting. The models obtained from the proposed method are also significantly faster than transistor-level models.
In this paper, for the first time, the deep gated recurrent unit (Deep GRU) is used as a new macromodeling approach for nonlinear circuits. Similar to Long Short-Term Memory (LSTM), the GRU has gating units that control the information flow and makes the network less prone to the vanishing gradient problem. Having a smaller number of gates causes GRU to have fewer parameters compared to LSTM leading to better model accuracy. Using the gates leads gradient formulations to have additive nature which helps them to be more resistant to vanishing and consequently learn long sequences of data. The proposed macromodeling method is capable of modeling nonlinear circuits more accurately and using fewer parameters compared to the conventional LSTM macromodeling method. To further improve the GRU performance, a regularization technique called Gaussian dropout is applied in this paper on deep GRU (GDGRU) to reduce the overfitting problem resulting in better test error. Additionally, the models obtained from the proposed techniques are remarkably faster than the original transistor-level models. To verify the superiority of the proposed method, time-domain modeling of three nonlinear circuits is provided. For these circuits, the comparisons of the accuracy and speed between the conventional recurrent neural network (RNN), the LSTM, and the proposed macromodeling methods are provided.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available