4.6 Article

RoBERTa-LSTM: A Hybrid Model for Sentiment Analysis With Transformer and Recurrent Neural Network

期刊

IEEE ACCESS
卷 10, 期 -, 页码 21517-21525

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2022.3152828

关键词

Sentiment analysis; Social networking (online); Training; Testing; Computational modeling; Analytical models; Blogs; Sentiment; transformer; RoBERTa; long short-term memory; LSTM; recurrent neural network; RNN

资金

  1. Fundamental Research Grant Scheme of the Ministry of Higher Education [FRGS/1/2019/ICT02/MMU/03/7]
  2. Multimedia University Internal Research Fund [MMUI/210112]

向作者/读者索取更多资源

Due to the rapid development of technology, social media has become increasingly common in people's daily lives. Sentiment analysis is an important technique to understand the sentiment behind texts. However, sentiment analysis faces challenges such as lexical diversity, imbalanced datasets, and long-distance dependencies. In this paper, a hybrid deep learning method is proposed that combines the strengths of sequence models and Transformer models. Experimental results show that this method outperforms state-of-the-art methods on different datasets.
Due to the rapid development of technology, social media has become more and more common in human daily life. Social media is a platform for people to express their feelings, feedback, and opinions. To understand the sentiment context of the text, sentiment analysis plays the role to determine whether the sentiment of the text is positive, negative, neutral or any other personal feeling. Sentiment analysis is prominent from the perspective of business or politics where it highly impacts the strategic decision making. The challenges of sentiment analysis are attributable to the lexical diversity, imbalanced dataset and long-distance dependencies of the texts. In view of this, a data augmentation technique with GloVe word embedding is leveraged to synthesize more lexically diverse samples by similar word vector replacements. The data augmentation also focuses on the oversampling of the minority classes to mitigate the imbalanced dataset problems. Apart from that, the existing sentiment analysis mostly leverages sequence models to encode the long-distance dependencies. Nevertheless, the sequence models require a longer execution time as the processing is done sequentially. On the other hand, the Transformer models require less computation time with parallelized processing. To that end, this paper proposes a hybrid deep learning method that combines the strengths of sequence model and Transformer model while suppressing the limitations of sequence model. Specifically, the proposed model integrates Robustly optimized BERT approach and Long Short-Term Memory for sentiment analysis. The Robustly optimized BERT approach maps the words into a compact meaningful word embedding space while the Long Short-Term Memory model captures the long-distance contextual semantics effectively. The experimental results demonstrate that the proposed hybrid model outshines the state-of-the-art methods by achieving F1-scores of 93%, 91%, and 90% on IMDb dataset, Twitter US Airline Sentiment dataset, and Sentiment140 dataset, respectively.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据