Neuron Specialization: Leveraging Intrinsic Task Modularity for Multilingual Machine Translation

Open Access
Authors
Publication date 2024
Host editors
  • Y. Al-Onaizan
  • M. Bansal
  • Y.-N. Chen
Book title The 2024 Conference on Empirical Methods in Natural Language Processing : Proceedings of the Conference
Book subtitle EMNLP 2024 : November 12-16, 2024
ISBN (electronic)
  • 9798891761643
Event 2024 Conference on Empirical Methods in Natural Language Processing
Pages (from-to) 6506-6527
Publisher Kerrville, TX: Association for Computational Linguistics
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Training a unified multilingual model promotes knowledge transfer but inevitably introduces negative interference. Language-specific modeling methods show promise in reducing interference. However, they often rely on heuristics to distribute capacity and struggle to foster cross-lingual transfer via isolated modules. In this paper, we explore intrinsic task modularity within multilingual networks and leverage these observations to circumvent interference under multilingual translation. We show that neurons in the feed-forward layers tend to be activated in a language-specific manner. Meanwhile, these specialized neurons exhibit structural overlaps that reflect language proximity, which progress across layers. Based on these findings, we propose Neuron Specialization, an approach that identifies specialized neurons to modularize feed-forward layers and then continuously updates them through sparse networks. Extensive experiments show that our approach achieves consistent performance gains over strong baselines with additional analyses demonstrating reduced interference and increased knowledge transfer.
Document type Conference contribution
Language English
Published at https://doi.org/10.18653/v1/2024.emnlp-main.374
Downloads
2024.emnlp-main.374 (Final published version)
Permalink to this page
Back