4.6 Article

Contrastive Learning Models for Sentence Representations

Publisher

ASSOC COMPUTING MACHINERY
DOI: 10.1145/3593590

Keywords

Sentence representation learning; contrastive learning; Data Augmentation; BERT

Ask authors/readers for more resources

Sentence representation learning is an important task in natural language processing, and the quality of learned representations directly impacts downstream tasks. Pretrained Transformer-based language models like BERT have shown moderate performance on various tasks. However, the anisotropy of BERT sentence embeddings hinders good results in semantic textual similarity tasks. Contrastive learning has been shown to alleviate this problem and improve representation performance. This article provides a summary and categorization of contrastive learning-based models for sentence representations, evaluation tasks, and future research directions, along with exhaustive experiments illustrating the quantitative improvement of various strategies on sentence representations.
Sentence representation learning is a crucial task in natural language processing, as the quality of learned representations directly influences downstream tasks, such as sentence classification and sentiment analysis. Transformer-based pretrained language models such as bidirectional encoder representations from transformers (BERT) have been extensively applied to various natural language processing tasks, and have exhibited moderately good performance. However, the anisotropy of the learned embedding space prevents BERT sentence embeddings from achieving good results in the semantic textual similarity tasks. It has been shown that contrastive learning can alleviate the anisotropy problem and significantly improve sentence representation performance. Therefore, there has been a surge in the development of models that utilize contrastive learning to fine-tune BERT-like pretrained language models to learn sentence representations. But no systematic review of contrastive learning models for sentence representations has been conducted. To fill this gap, this article summarizes and categorizes the contrastive learning based sentence representation models, common evaluation tasks for assessing the quality of learned representations, and future research directions. Furthermore, we select several representative models for exhaustive experiments to illustrate the quantitative improvement of various strategies on sentence representations.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available