4.6 Article

VSCA: A Sentence Matching Model Incorporating Visual Perception

期刊

COGNITIVE COMPUTATION
卷 15, 期 1, 页码 323-336

出版社

SPRINGER
DOI: 10.1007/s12559-022-10074-8

关键词

Natural language processing; Sentence matching; Variational autoencoder; Spatial attention

向作者/读者索取更多资源

This study proposes a novel sentence matching model (VSCA) that utilizes a variational autoencoder (VAE)-based attention mechanism to construct basic attention feature maps and combines spatial attention mechanism with visual perception to capture multilevel semantic information. Experimental results demonstrate that VSCA outperforms pretrained models like BERT on the LCQMC dataset and performs well on the PAWS-X data. In addition, VSCA has lower time and space complexity while capturing rich attentional information and detailed information.
Stacking multiple layers of attention networks can significantly improve a model's performance. However, this also increases the model's time and space complexity, making it difficult for the model to capture detailed information on the underlying features. We propose a novel sentence matching model (VSCA) that uses a new attention mechanism based on variational autoencoders (VAE), which exploits the contextual information in sentences to construct a basic attention feature map and combines it with VAE to generate multiple sets of related attention feature maps for fusion. Furthermore, VSCA introduces a spatial attention mechanism that combines visual perception to capture multilevel semantic information. The experimental results show that our proposed model outperforms pretrained models such as BERT on the LCQMC dataset and performs well on the PAWS-X data. Our work consists of two parts. The first part compares the proposed sentence matching model with state-of-the-art pretrained models such as BERT. The second part conducts innovative research on applying VAE and spatial attention mechanisms in NLP. The experimental results on the related datasets show that the proposed method has satisfactory performance, and VSCA can capture rich attentional information and detailed information with less time and space complexity. This work provides insights into the application of VAE and spatial attention mechanisms in NLP.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据