Towards Fast Multilingual LLM Inference: Speculative Decoding and Specialized Drafters

Euiin Yi, Taehyeon Kim, Hongseok Jeung, Du-Seong Chang, Se-Young Yun


Abstract
Large language models (LLMs) have revolutionized natural language processing and broadened their applicability across diverse commercial applications. However, the deployment of these models is constrained by high inference time in multilingual settings. To mitigate this challenge, this paper explores a training recipe of an assistant model in speculative decoding, which are leveraged to draft and-then its future tokens are verified by the target LLM. We show that language-specific draft models, optimized through a targeted pretrain-and-finetune strategy, substantially brings a speedup of inference time compared to the previous methods. We validate these models across various languages in inference time, out-of-domain speedup, and GPT-4o evaluation.
Anthology ID:
2024.emnlp-main.602
Volume:
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2024
Address:
Miami, Florida, USA
Editors:
Yaser Al-Onaizan, Mohit Bansal, Yun-Nung Chen
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
10789–10802
Language:
URL:
https://aclanthology.org/2024.emnlp-main.602
DOI:
Bibkey:
Cite (ACL):
Euiin Yi, Taehyeon Kim, Hongseok Jeung, Du-Seong Chang, and Se-Young Yun. 2024. Towards Fast Multilingual LLM Inference: Speculative Decoding and Specialized Drafters. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pages 10789–10802, Miami, Florida, USA. Association for Computational Linguistics.
Cite (Informal):
Towards Fast Multilingual LLM Inference: Speculative Decoding and Specialized Drafters (Yi et al., EMNLP 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.emnlp-main.602.pdf
Software:
 2024.emnlp-main.602.software.zip