Capturing Perspectives of Crowdsourced Annotators in Subjective Learning Tasks

Open Access
Authors
  • N. Mokhberian
  • M.G. Marmarelis
  • F.R. Hopp
  • V. Basile
  • F. Morstatter
  • K. Lerman
Publication date 2024
Host editors
  • K. Duh
  • H. Gomez
  • S. Bethard
Book title The 2024 Conference of the North American Chapter of the Association for Computational Linguistics : proceedings of the conference
Book subtitle NAACL 2024 : June 16-21, 2024
ISBN (electronic)
  • 9798891761148
Event 2024 Conference of the North American Chapter of the Association for Computational Linguistics
Volume | Issue number 1
Pages (from-to) 7337-7349
Publisher Kerrville, TX: Association for Computational Linguistics
Organisations
  • Faculty of Social and Behavioural Sciences (FMG) - Amsterdam School of Communication Research (ASCoR)
Abstract
Supervised classification heavily depends on datasets annotated by humans. However, in subjective tasks such as toxicity classification, these annotations often exhibit low agreement among raters. Annotations have commonly been aggregated by employing methods like majority voting to determine a single ground truth label. In subjective tasks, aggregating labels will result in biased labeling and, consequently, biased models that can overlook minority opinions. Previous studies have shed light on the pitfalls of label aggregation and have introduced a handful of practical approaches to tackle this issue. Recently proposed multi-annotator models, which predict labels individually per annotator, are vulnerable to under-determination for annotators with few samples. This problem is exacerbated in crowdsourced datasets. In this work, we propose Annotator Aware Representations for Texts (AART) for subjective classification tasks. Our approach involves learning representations of annotators, allowing for exploration of annotation behaviors. We show the improvement of our method on metrics that assess the performance on capturing individual annotators’ perspectives. Additionally, we demonstrate fairness metrics to evaluate our model’s equability of performance for marginalized annotators compared to others.
Document type Conference contribution
Language English
Published at https://doi.org/10.18653/v1/2024.naacl-long.407
Other links https://aclanthology.org/2024.naacl-long.407.mp4
Downloads
Permalink to this page
Back