Representational Isomorphism and Alignment of Multilingual Large Language Models

Open Access
Authors
Publication date 2024
Host editors
  • Y. Al-Onaizan
  • M. Bansal
  • Y.-N. Chen
Book title The 2024 Conference on Empirical Methods in Natural Language Processing : Findings of EMNLP 2024
Book subtitle EMNLP 2024 : November 12-16, 2024
ISBN (electronic)
  • 9798891761681
Event 2024 Conference on Empirical Methods in Natural Language Processing
Pages (from-to) 14074-14085
Number of pages 12
Publisher Kerrville, TX: Association for Computational Linguistics
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
In this paper, we investigate the capability of Large Language Models (LLMs) to represent texts in multilingual contexts. Our findings show that sentence representations derived from LLMs exhibit a high degree of isomorphism across languages.This existing isomorphism can facilitate representational alignments in zero-shot and few-shot settings.Specifically, by applying a contrastive objective at the representation level with only a small number of translation pairs (e.g., 100), we substantially improve models’ performance on Semantic Textual Similarity (STS) tasks across languages. This representation-level approach proves to be more efficient and effective for semantic alignment than continued pretraining or instruction tuning. Interestingly, we also observe substantial STS improvements within individual languages, even without a monolingual objective specifically designed for this purpose.
Document type Conference contribution
Language English
Related publication Representational Isomorphism and Alignment of Multilingual Large Language Models
Published at https://doi.org/10.18653/v1/2024.findings-emnlp.823
Downloads
2024.findings-emnlp.823 (Final published version)
Permalink to this page
Back