4.7 Article

A New Macromodeling Method Based on Deep Gated Recurrent Unit Regularized With Gaussian Dropout for Nonlinear Circuits

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TCSI.2023.3264616

关键词

Integrated circuit modeling; Solid modeling; Recurrent neural networks; Logic gates; Neural networks; Nonlinear circuits; Training; Computer-aided design (CAD); deep neural network; gated recurrent unit (GRU); Gaussian dropout; macromodeling; nonlinear circuits; recurrent neural network (RNN)

向作者/读者索取更多资源

This paper introduces the use of deep gated recurrent unit (Deep GRU) as a novel macromodeling approach for nonlinear circuits. The GRU, similar to LSTM, has gating units that control information flow and address the vanishing gradient problem. The proposed method outperforms the conventional LSTM macromodeling method in terms of accuracy and parameter efficiency, and the application of Gaussian dropout on deep GRU further improves its performance and reduces overfitting. The models obtained from the proposed method are also significantly faster than transistor-level models.
In this paper, for the first time, the deep gated recurrent unit (Deep GRU) is used as a new macromodeling approach for nonlinear circuits. Similar to Long Short-Term Memory (LSTM), the GRU has gating units that control the information flow and makes the network less prone to the vanishing gradient problem. Having a smaller number of gates causes GRU to have fewer parameters compared to LSTM leading to better model accuracy. Using the gates leads gradient formulations to have additive nature which helps them to be more resistant to vanishing and consequently learn long sequences of data. The proposed macromodeling method is capable of modeling nonlinear circuits more accurately and using fewer parameters compared to the conventional LSTM macromodeling method. To further improve the GRU performance, a regularization technique called Gaussian dropout is applied in this paper on deep GRU (GDGRU) to reduce the overfitting problem resulting in better test error. Additionally, the models obtained from the proposed techniques are remarkably faster than the original transistor-level models. To verify the superiority of the proposed method, time-domain modeling of three nonlinear circuits is provided. For these circuits, the comparisons of the accuracy and speed between the conventional recurrent neural network (RNN), the LSTM, and the proposed macromodeling methods are provided.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据