Low-Rank Updates of pre-trained Weights for Multi-Task Learning

Alexandre Audibert, Massih R Amini, Konstantin Usevich, Marianne Clausel


Abstract
Multi-Task Learning used with pre-trained models has been quite popular in the field of Natural Language Processing in recent years. This framework remains still challenging due to the complexity of the tasks and the challenges associated with fine-tuning large pre-trained models. In this paper, we propose a new approach for Multi-task learning which is based on stacking the weights of Neural Networks as a tensor. We show that low-rank updates in the canonical polyadic tensor decomposition of this tensor of weights lead to a simple, yet efficient algorithm, which without loss of performance allows to reduce considerably the model parameters. We investigate the interactions between tasks inside the model as well as the inclusion of sparsity to find the best tensor rank and to increase the compression rate. Our strategy is consistent with recent efforts that attempt to use constraints to fine-tune some model components. More precisely, we achieve equivalent performance as the state-of-the-art on the General Language Understanding Evaluation benchmark by training only 0.3 of the parameters per task while not modifying the baseline weights.
Anthology ID:
2023.findings-acl.476
Volume:
Findings of the Association for Computational Linguistics: ACL 2023
Month:
July
Year:
2023
Address:
Toronto, Canada
Editors:
Anna Rogers, Jordan Boyd-Graber, Naoaki Okazaki
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
7544–7554
Language:
URL:
https://aclanthology.org/2023.findings-acl.476
DOI:
10.18653/v1/2023.findings-acl.476
Bibkey:
Cite (ACL):
Alexandre Audibert, Massih R Amini, Konstantin Usevich, and Marianne Clausel. 2023. Low-Rank Updates of pre-trained Weights for Multi-Task Learning. In Findings of the Association for Computational Linguistics: ACL 2023, pages 7544–7554, Toronto, Canada. Association for Computational Linguistics.
Cite (Informal):
Low-Rank Updates of pre-trained Weights for Multi-Task Learning (Audibert et al., Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-acl.476.pdf