Interpreting Predictive Probabilities Model Confidence or Human Label Variation?

Open Access
Authors
Publication date 2024
Host editors
  • Y. Graham
  • M. Purver
Book title The 18th Conference of the European Chapter of the Association for Computational Linguistics
Book subtitle proceedings of the conference : EACL 2024 : March 17-22, 2024
ISBN (electronic)
  • 9798891760899
Event 18th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2024
Volume | Issue number 2
Pages (from-to) 268-277
Number of pages 10
Publisher Kerrville, TX: Association for Computational Linguistics
Organisations
  • Faculty of Science (FNWI)
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract

With the rise of increasingly powerful and user-facing NLP systems, there is growing interest in assessing whether they have a good representation of uncertainty by evaluating the quality of their predictive distribution over outcomes. We identify two main perspectives that drive starkly different evaluation protocols. The first treats predictive probability as an indication of model confidence; the second as an indication of human label variation. We discuss their merits and limitations, and take the position that both are crucial for trustworthy and fair NLP systems, but that exploiting a single predictive distribution is limiting. We recommend tools and highlight exciting directions towards models with disentangled representations of uncertainty about predictions and uncertainty about human labels.

Document type Conference contribution
Note With supplementary video
Language English
Published at https://doi.org/10.18653/v1/2024.eacl-short.24
Other links https://www.scopus.com/pages/publications/85189941643
Downloads
2024.eacl-short.24 (Final published version)
Supplementary materials
Permalink to this page
Back