Task-Incremental Learning on Long Text Sequences

Natalia Graziuso, Andrea Zugarini, Stefano Melacci


Abstract
The extraordinary results achieved by Large Language Models are paired with issues that are critical in real-world applications. The costs of inference and, in particular, training are extremely large, both in terms of time and computational resources, and they become prohibitive when working in dynamic environments, where data and tasks are progressively provided over time. The model must be able to adapt to new knowledge, new domains, new settings, without forgetting the previously learned skills. Retraining from scratch easily becomes too costly, thus Continual Learning strategies are of crucial importance. This is even more evident when data consist of “long” documents, that require several resources to be processed by modern neural models, leading to very long prompts. This paper investigates LLM-based Task-Incremental Learning in the case of tasks exploiting long sequences of text, as it is typical in summarization, question-answering on long documents, reviewing long contracts, and several others. We show how adapting the model by Task Arithmetic with LoRA, which was proposed for visual data, yields promising results also in the case of such “long” text data. To our best knowledge, this is the first work along this challenging direction. The outcome of the investigation of this paper is generic enough to represent an important starting point for further research in processing linguistic data in every language.
Anthology ID:
2024.clicit-1.49
Volume:
Proceedings of the 10th Italian Conference on Computational Linguistics (CLiC-it 2024)
Month:
December
Year:
2024
Address:
Pisa, Italy
Editors:
Felice Dell'Orletta, Alessandro Lenci, Simonetta Montemagni, Rachele Sprugnoli
Venue:
CLiC-it
SIG:
Publisher:
CEUR Workshop Proceedings
Note:
Pages:
410–416
Language:
URL:
https://aclanthology.org/2024.clicit-1.49/
DOI:
Bibkey:
Cite (ACL):
Natalia Graziuso, Andrea Zugarini, and Stefano Melacci. 2024. Task-Incremental Learning on Long Text Sequences. In Proceedings of the 10th Italian Conference on Computational Linguistics (CLiC-it 2024), pages 410–416, Pisa, Italy. CEUR Workshop Proceedings.
Cite (Informal):
Task-Incremental Learning on Long Text Sequences (Graziuso et al., CLiC-it 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.clicit-1.49.pdf