Whose truth is it anyway? An experiment on annotation bias in times of factual opinion polarization
| Authors |
|
|---|---|
| Publication date | 2025 |
| Journal | Communication Methods and Measures |
| Volume | Issue number | 19 | 4 |
| Pages (from-to) | 332-349 |
| Organisations |
|
| Abstract |
Information shapes citizens’ political decision-making. This process is amply studied by social scientists, for whom human annotation is a crucial instrument in their toolkit. Due to the democratization of data and the advances in NLP more data can be analyzed or classified, making these benchmarks more important than ever: An algorithm trained on biased data will reproduce and often exacerbate bias. Currently, disagreement and bias are often conflated–ignoring the possibility of annotator sample bias and valid disagreement. In a preregistered experiment in the Netherlands and a close replication in the U.S. we show that personal characteristics of annotators, like political ideology or knowledge, can interfere with annotators’ judgment of political stances. Our results show that to improve annotated data for automated text analyses, and for stance detection models in particular, we need to critically evaluate how we create our gold standards.
|
| Document type | Article |
| Language | English |
| Published at | https://doi.org/10.1080/19312458.2025.2562034 |
| Other links | https://www.scopus.com/pages/publications/105017044553 |
| Downloads |
Whose truth is it anyway?
(Final published version)
|
| Permalink to this page | |
