Misery loves complexity : exploring linguistic complexity in the context of emotion detection

Publication type
C1
Publication status
Published
Authors
Singh, P., De Bruyne, L., De Clercq, O., & Lefever, E.
Editor
Houda Bouamor, Juan Pino and Kalika Bali
Series
Findings of the Association for Computational Linguistics : EMNLP 2023
Pagination
12871-12880
Publisher
Association for Computational Linguistics
Conference
2023 Conference on Empirical Methods in Natural Language Processing Singapore (EMNLP 2023) (Singapore)
Download
(.pdf)
View in Biblio
(externe link)

Abstract

Given the omnipresence of social media in our society, thoughts and opinions are being shared online in an unprecedented manner. This means that both positive and negative emotions can be equally and freely expressed. However, the negativity bias posits that human beings are inherently drawn to and more moved by negativity and, as a consequence, negative emotions get more traffic. Correspondingly, when writing about emotions this negativity bias could lead to expressions of negative emotions that are linguistically more complex. In this paper, we attempt to use readability and linguistic complexity metrics to better understand the manifestation of emotions on social media platforms like Reddit based on the widely-used GoEmotions dataset. We demonstrate that according to most metrics, negative emotions indeed tend to generate more complex text than positive emotions. In addition, we examine whether a higher complexity hampers the automatic identification of emotions. To answer this question, we fine-tuned three state-of-the-art transformers (BERT, RoBERTa, and SpanBERT) on the same emotion detection dataset. We demonstrate that these models often fail to predict emotions for the more complex texts. More advanced LLMs like RoBERTa and SpanBERT also fail to improve by significant margins on complex samples. This calls for a more nuanced interpretation of the emotion detection performance of transformer models. We make the automatically annotated data available for further research at: https://huggingface.co/datasets/pranaydeeps/CAMEO