期刊
KNOWLEDGE-BASED SYSTEMS
卷 227, 期 -, 页码 -出版社
ELSEVIER
DOI: 10.1016/j.knosys.2021.107220
关键词
Sentiment knowledge graph; BERT; Aspect-sentiment analysis
资金
- Basic Scientific Research Projects of Wenzhou, China [G2020024]
The knowledge-enabled language representation model BERT proposed in this work enhances aspect-based sentiment analysis by injecting domain knowledge and leveraging an external sentiment knowledge graph, resulting in more accurate and explainable results.
To provide explainable and accurate aspect terms and the corresponding aspect-sentiment detection, it is often useful to take external domain-specific knowledge into consideration. In this work, we propose a knowledge-enabled language representation model BERT for aspect-based sentiment analysis. Specifically, our proposal leverages the additional information from a sentiment knowledge graph by injecting sentiment domain knowledge into the language representation model, which obtains the embedding vectors of entities in the sentiment knowledge graph and words in the text in a consistent vector space. In addition, the model is capable of achieving better performance with a small amount of training data by incorporating external domain knowledge into the language representation model to compensate for the limited training data. As a result, our model is able to provide explainable and detailed results for aspect-based sentiment analysis. Experimental results demonstrate the effectiveness of the proposed method, showing that the knowledge-enabled BERT is an excellent choice for solving aspect-based sentiment analysis problems. (C) 2021 Elsevier B.V. All rights reserved.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据