Human-like Linguistic Biases in Neural Speech Models: Phonetic Categorization and Phonotactic Constraints in Wav2Vec2.0

Open Access
Authors
Publication date 2024
Journal Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
Volume | Issue number 25
Pages (from-to) 4593-4597
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
What do deep neural speech models know about phonology? Existing work has examined the encoding of individual linguistic units such as phonemes in these models. Here we investigate interactions between units. Inspired by classic experiments on human speech perception, we study how Wav2Vec2 resolves phonotactic constraints. We synthesize sounds on an acoustic continuum between /l/ and /r/ and embed them in controlled contexts where only /l/, only /r/, or neither occur in English. Like humans, Wav2Vec2 models show a bias towards the phonotactically admissable category in processing such ambiguous sounds. Using simple measures to analyze model internals on the level of individual stimuli, we find that this bias emerges in early layers of the model's Transformer module. This effect is amplified by ASR finetuning but also present in fully self-supervised models. Our approach demonstrates how controlled stimulus designs can help localize specific linguistic knowledge in neural speech models.
Document type Article
Language English
Published at https://doi.org/10.21437/Interspeech.2024-2490
Other links https://www.proceedings.com/76640.html
Downloads
deheerkloots24_interspeech (Final published version)
Permalink to this page
Back