Stepmothers are mean and academics are pretentious: What do pretrained language models learn about you?'
| Authors | |
|---|---|
| Publication date | 2021 |
| Host editors |
|
| Book title | 2021 Conference on Empirical Methods in Natural Language Processing |
| Book subtitle | EMNLP 2021 : proceedings of the conference : November 7-11, 2021 |
| ISBN (electronic) |
|
| Event | 2021 Conference on Empirical Methods in Natural Language Processing |
| Pages (from-to) | 1477-1491 |
| Publisher | Stroudsburg, PA: The Association for Computational Linguistics |
| Organisations |
|
| Abstract |
In this paper, we investigate what types of stereotypical information are captured by pretrained language models. We present the first dataset comprising stereotypical attributes of a range of social groups and propose a method to elicit stereotypes encoded by pretrained language models in an unsupervised fashion. Moreover, we link the emergent stereotypes to their manifestation as basic emotions as a means to study their emotional effects in a more generalized manner. To demonstrate how our methods can be used to analyze emotion and stereotype shifts due to linguistic experience, we use fine-tuning on news sources as a case study. Our experiments expose how attitudes towards different social groups vary across models and how quickly emotions and stereotypes can shift at the fine-tuning stage.
|
| Document type | Conference contribution |
| Note | With supplementary video |
| Language | English |
| Published at | https://doi.org/10.18653/v1/2021.emnlp-main.111 |
| Downloads |
2021.emnlp-main.111
(Final published version)
|
| Supplementary materials | |
| Permalink to this page | |