4.7 Article

Scalable multi-channel dilated CNN-BiLSTM model with attention mechanism for Chinese textual sentiment analysis

出版社

ELSEVIER
DOI: 10.1016/j.future.2021.01.024

关键词

Chinese textual sentiment analysis; Dilated convolutional neural network; Bidirectional long short-term memory; Attention mechanism; Scalable multi-channel

资金

  1. Natural Science Foundation of China [61702066, 11747125]
  2. Major Project of Science and Technology Research Program of Chongqing Education Commission of China [KJZD-M201900601]
  3. Chongqing Research Program of Basic Research and Frontier Technology, China [cstc2017jcyjAX0256, cstc2018jcyjAX0154]
  4. Chongqing Municipal Key Laboratory of Institutions of Higher Education, China [cqupt-mct-201901]
  5. Chongqing Key Laboratory of Mobile Communications Technology, China [cqupt-mct-202002]
  6. Engineering Research Center of Mobile Communications, Ministry of Education [cqupt-mct-202006]
  7. Research Innovation Program for Postgraduate of Chongqing, China [CYS17217, CYS18238]

向作者/读者索取更多资源

This paper proposes a scalable multi-channel dilated joint architecture of convolutional neural network and bidirectional long short-term memory model with an attention mechanism to analyze the sentiment tendency of Chinese texts. The model can extract both the original context features and the multiscale high-level context features, and utilize the attention mechanism to further distinguish the difference of features. Additionally, an adaptive weighted loss function is designed to effectively avoid the imbalance of classes in training data.
Due to the complex semantics of natural language, the multi-sentiment polarity of words, and the long-dependence of sentiments between words, the existing sentiment analysis methods (especially Chinese textual sentiment analysis) still face severe challenges. Aware of these issues, this paper proposes a scalable multi-channel dilated joint architecture of convolutional neural network and bidirectional long short-term memory (CNN-BiLSTM) model with an attention mechanism to analyze the sentiment tendency of Chinese texts. Through the multi-channel structure, this model can extract both the original context features and the multiscale high-level context features. Importantly, the number of the model channel can be optimally expanded according to the actual corpus. Furthermore, the attention mechanism including local attention and global attention is adopted to further distinguish the difference of features. The former is employed to weight the output features of each channel, and the latter is used to weight the fused features of all channels. Besides, an adaptive weighted loss function is designed to effectively avoid the imbalance of classes in training data. Finally, several experiments are performed to demonstrate the superior performance of the proposed model on two public datasets. Compared with word-level methods, the accuracy and Macro-F1 are respectively increased by over 1.19% and 0.9% on NLPCC2017-ECGC corpus, the accuracy and F1 are respectively increased by more than 1.7% and 1.214% on ChnSentiCorp-Htl-unba-10000 corpus. Compared with char-level pre-training methods, the accuracy and Macro-F1 also respectively achieve an improvement of over 3.416% and 4.324% on the NLPCC2017-ECGC corpus, the accuracy and F1 are respectively increased by more than 0.14% and 3% on the ChnSentiCorp-Htl-unba-10000 corpus. (C) 2021 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据