4.7 Article

AOBERT: All-modalities-in-One BERT for multimodal sentiment analysis

Journal

INFORMATION FUSION
Volume 92, Issue -, Pages 37-45

Publisher

ELSEVIER
DOI: 10.1016/j.inffus.2022.11.022

Keywords

Multimodal Sentiment Analysis; Single -stream Transformer; Multimodal Masked Language Model; Alignment Prediction

Ask authors/readers for more resources

Multimodal sentiment analysis uses various modalities to predict sentiment, but traditional methods suffer from loss of intramodality and inter-modality. AOBERT, a single-stream transformer pre-trained on two tasks, achieves state-of-the-art results and addresses this problem effectively.
Multimodal sentiment analysis utilizes various modalities such as Text, Vision and Speech to predict sentiment. As these modalities have unique characteristics, methods have been developed for fusing features. However, the overall modality characteristics are not guaranteed, because traditional fusion methods have some loss of intramodality and inter-modality. To solve this problem, we introduce a single-stream transformer, All-modalities-inOne BERT (AOBERT). The model is pre-trained on two tasks simultaneously: Multimodal Masked Language Modeling (MMLM) and Alignment Prediction (AP). The dependency and relationship between modalities can be determined using two pre-training tasks. AOBERT achieved state-of-the-art results on the CMU-MOSI, CMUMOSEI, and UR-FUNNY datasets. Furthermore, ablation studies that validated combinations of modalities, effects of MMLM and AP and fusion methods confirmed the effectiveness of the proposed model.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available