PILLOW: Enhancing Efficient Instruction Fine-tuning via Prompt Matching

Zhenting Qi, Xiaoyu Tan, Shaojie Shi, Chao Qu, Yinghui Xu, Yuan Qi


Abstract
Instruction fine-tuning has conventionally been employed to adapt Large Language Models (LLMs) to a variety of diverse tasks. Nonetheless, this technique often necessitates substantial computational resources, making it impractical for deployment by individuals or small-scale entities. Recently, Low-Rank Adaptation (LoRA) has become a promising alternative, offering tuning capabilities with reduced resource overhead. However, attaining satisfactory performance through the fine-tuning of LoRA is a non-trivial challenge. In this paper, we propose PILLOW, which aims to improve LoRA’s performance by leveraging LLM’s in-context learning capability through prompt matching via reinforcement learning in resource-constrained environments. Specifically, PILLOW incorporates a matching network that selects prompts from a user-defined pool, concatenates the optimal prompts given the user instruction, and performs inference using the LoRA-fine-tuned LLMs. Compared with typical instruction fine-tuning methods, PILLOW exhibits commensurate performance on various evaluation metrics, utilizing only consumer-grade GPU resources and exhibiting a large increase in training efficiency.
Anthology ID:
2023.emnlp-industry.45
Volume:
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Industry Track
Month:
December
Year:
2023
Address:
Singapore
Editors:
Mingxuan Wang, Imed Zitouni
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
471–482
Language:
URL:
https://aclanthology.org/2023.emnlp-industry.45
DOI:
10.18653/v1/2023.emnlp-industry.45
Bibkey:
Cite (ACL):
Zhenting Qi, Xiaoyu Tan, Shaojie Shi, Chao Qu, Yinghui Xu, and Yuan Qi. 2023. PILLOW: Enhancing Efficient Instruction Fine-tuning via Prompt Matching. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: Industry Track, pages 471–482, Singapore. Association for Computational Linguistics.
Cite (Informal):
PILLOW: Enhancing Efficient Instruction Fine-tuning via Prompt Matching (Qi et al., EMNLP 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.emnlp-industry.45.pdf
Video:
 https://aclanthology.org/2023.emnlp-industry.45.mp4