An Empirical Study on the Transferability of Transformer Modules in Parameter-Efficient Fine-Tuning
| Authors |
|
|---|---|
| Publication date | 2022 |
| Host editors |
|
| Book title | Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing |
| Book subtitle | December 7-11, 2022, Abu Dhabi, United Arab Emirates |
| Event | The 2022 Conference on Empirical Methods in Natural Language Processing |
| Pages (from-to) | 10617–10625 |
| Publisher | Association for Computational Linguistics |
| Organisations |
|
| Abstract |
Parameter-efficient fine-tuning approaches have recently garnered a lot of attention. Having considerably lower number of train-able weights, these methods can bring about scalability and computational effectiveness. In this paper, we look for optimal sub-networks and investigate the capability of different transformer modules in transferring knowledge from a pre-trained model to a downstream task. Our empirical results suggest that every transformer module in BERT can act as a winning ticket: fine-tuning each specific module while keeping the rest of the network frozen can lead to comparable performance to the full fine-tuning. Among different modules, LayerNorms exhibit the best capacity for knowledge transfer with limited trainable weights, to the extent that, with only 0.003% of all parameters in the layer-wise analysis, they show acceptable performance on various target tasks. On the reasons behind their effectiveness, we argue that their notable performance could be attributed to their high-magnitude weights compared to that of the other modules in the pre-trained BERT. The code for this paper is freely available at https://github.com/m-tajari/transformer-transferability.
|
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.18653/v1/2022.emnlp-main.726 |
| Downloads |
2022.emnlp-main.726
(Final published version)
|
| Permalink to this page | |