Another approach to agreement measurement and prediction with emotion annotations

Publication type
C1
Publication status
In press
Authors
Du, Q, & Hoste, V.
Series
Proceedings of the 19th Linguistic Annotation Workshop (LAW-XIX)
Pagination
1-16
Publisher
Association for Computational Linguistics (ACL)
Conference
19th Linguistic Annotation Workshop Co-located with ACL 2025 (Vienna, Austria)
Download
(.pdf)
View in Biblio
(externe link)

Abstract

Emotion annotation, as an inherently subjective task, often suffers from significant interannotator disagreement when evaluated using traditional metrics like kappa or alpha. These metrics often fall short of capturing the nuanced nature of disagreement, especially in multimodal settings. This study introduces Absolute
Annotation Difference (AAD), a novel metric offering a complementary perspective on inter- and intra-annotator agreement across different modalities. Our analysis reveals that AAD not only identifies overall agreement levels but also uncovers fine-grained disagreement patterns across modalities often overlooked by conventional metrics. Furthermore, we propose an AAD-based RMSE variant for predicting annotation disagreement. Through extensive experiments on the large-scale DynaSent corpus, we demonstrate that our approach significantly
improves disagreement prediction accuracy, rising from 41.71% to 51.64% and outperforming existing methods. Cross-dataset prediction results suggest good generalization. These findings underscore AAD’s potential to
enhance annotation agreement analysis and provide deeper insights into subjective NLP tasks. Future work will investigate its applicability to broader emotion-related tasks and other subjective annotation scenarios.