These are not the Stereotypes You are Looking For: Bias and Fairness in Authorial Gender Attribution

Open Access
Authors
Publication date 2017
Host editors
  • D. Hovy
  • S. Spruit
  • M. Mitchell
  • E.M. Bender
  • M. Strube
  • H. Wallach
Book title Ethics in Natural Language Processing
Book subtitle EACL 2017 : Proceedings of the First ACL Workshop : april 4th, 2017, Valencia, Spain
ISBN (electronic)
  • 9781945626470
Event Ethics in Natural Language Processing
Pages (from-to) 12-22
Publisher Stroudsburg, PA: Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
Stylometric and text categorization results show that author gender can be discerned in texts with relatively high accuracy. However, it is difficult to explain what gives rise to these results and there are many possible confounding factors, such as the domain, genre, and target audience of a text. More fundamentally, such classification efforts risk invoking stereotyping and essentialism. We explore this issue in two datasets of Dutch literary novels, using commonly used descriptive (LIWC, topic modeling) and predictive (machine learning) methods. Our results show the importance of controlling for variables in the corpus and we argue for taking care not to overgeneralize from the results.
Document type Conference contribution
Language English
Published at https://doi.org/10.18653/v1/W17-1602
Downloads
koolenvancranenburgh2017thesearenot (Final published version)
Permalink to this page
Back