4.6 Article

Visual-Textual Sentiment Analysis Enhanced by Hierarchical Cross-Modality Interaction

期刊

IEEE SYSTEMS JOURNAL
卷 15, 期 3, 页码 4303-4314

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JSYST.2020.3026879

关键词

Sentiment analysis; Semantics; Visualization; Analytical models; Task analysis; Convolutional neural networks; Learning systems; Attention mechanism; multimodal convolutional neural networks; sentiment analysis; transfer learning

资金

  1. National Natural Science Foundation of China [61772133, 61972087]
  2. National Social Science Foundation of China [19@ZH014]
  3. Jiangsu Provincial Key Project [BE2018706]
  4. Natural Science Foundation of Jiangsu province [SBK2019022870]
  5. Jiangsu Provincial Key Laboratory of Computer Networking Technology
  6. Jiangsu Provincial Key Laboratory of Network and Information Security [BM2003201]
  7. Key Laboratory of Computer Network and Information Integration of Ministry of Education of China [93K-9]

向作者/读者索取更多资源

This article proposed a hierarchical cross-modality interaction model for visual-textual sentiment analysis, emphasizing consistency and correlation across modalities and addressing noise and joint understanding issues. Through experiments, the framework outperformed existing methods, with phrase-level text fragments playing an important role in joint visual-textual sentiment analysis.
Visual-textual sentiment analysis could benefit user understanding in online social networks and enable many useful applications like user profiling and recommendation. However, it faces a set of new challenges, i.e., exacerbated noise problem caused by irrelevant or redundant information in different modalities, and the gap in joint understanding for multimodal sentiment. In this article, we propose hierarchical cross-modality interaction model for visual-textual sentiment analysis. Our model emphasises the consistency and correlation across modalities, by extracting the semantic and sentiment interactions between image and text in a hierarchical way, which could cope with the noise and joint understanding issues, respectively. Hierarchical attention mechanism is first adopted to capture the semantic interaction and purify the information in one modality with the help of the other. Then, multimodal convolutional neural network, which could fully exploit cross-modality sentiment interaction is incorporated, and better joint visual-textual representation is generated. A transfer learning method is further designed to alleviate the impact of noises in real social data. Through extensive experiments on two datasets, we show that our proposed framework greatly surpasses the state-of-the-art approaches. Specifically, the phrase-level text fragment plays an important role in interacting with image regions for joint visual-textual sentiment analysis.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据