4.7 Article

The Biases of Pre-Trained Language Models: An Empirical Study on Prompt-Based Sentiment Analysis and Emotion Detection

期刊

IEEE TRANSACTIONS ON AFFECTIVE COMPUTING
卷 14, 期 3, 页码 1743-1753

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TAFFC.2022.3204972

关键词

Task analysis; Emotion recognition; Sentiment analysis; Computational modeling; Affective computing; Taxonomy; Analytical models; Emotion detection; pre-trained language model; prompt; sentiment analysis

向作者/读者索取更多资源

With the breakthrough of large-scale pre-trained language model (PLM) technology, prompt-based classification tasks, such as sentiment analysis and emotion detection, have gained increasing attention. This study conducts a systematic empirical study on prompt-based sentiment analysis and emotion detection to investigate the biases of PLMs in affective computing.
Thanks to the breakthrough of large-scale pre-trained language model (PLM) technology, prompt-based classification tasks, e.g., sentiment analysis and emotion detection, have raised increasing attention. Such tasks are formalized as masked language prediction tasks which are in line with the pre-training objects of most language models. Thus, one can use a PLM to infer the masked words in a downstream task, then obtaining label predictions with manually defined label-word mapping templates. Prompt-based affective computing takes the advantages of both neural network modeling and explainable symbolic representations. However, there still remain many unclear issues related to the mechanisms of PLMs and prompt-based classification. We conduct a systematic empirical study on prompt-based sentiment analysis and emotion detection to study the biases of PLMs towards affective computing. We find that PLMs are biased in sentiment analysis and emotion detection tasks with respect to the number of label classes, emotional label-word selections, prompt templates and positions, and the word forms of emotion lexicons.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据