Journal
MATHEMATICS
Volume 11, Issue 10, Pages -Publisher
MDPI
DOI: 10.3390/math11102335
Keywords
document-level multimodal sentiment classification; graph convolutional networks
Categories
Ask authors/readers for more resources
An increasing number of people use different modalities to convey their opinions. In this work, we propose a novel Multimodal Interactive and Fusion Graph Convolutional Network for document-level multimodal sentiment analysis. The model incorporates text, images, and image captions to enhance semantics delivery and filter visual noise. Experimental results on a multimodal dataset show satisfying performance in sentiment analysis tasks.
An increasing number of people tend to convey their opinions in different modalities. For the purpose of opinion mining, sentiment classification based on multimodal data becomes a major focus. In this work, we propose a novel Multimodal Interactive and Fusion Graph Convolutional Network to deal with both texts and images on the task of document-level multimodal sentiment analysis. The image caption is introduced as an auxiliary, which is aligned with the image to enhance the semantics delivery. Then, a graph is constructed with the sentences and images generated as nodes. In line with the graph learning, the long-distance dependencies can be captured while the visual noise can be filtered. Specifically, a cross-modal graph convolutional network is built for multimodal information fusion. Extensive experiments are conducted on a multimodal dataset from Yelp. Experimental results reveal that our model obtains a satisfying working performance in DLMSA tasks.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available