Flow Matching for Conditional Text Generation in a Few Sampling Steps

Vincent Hu, Di Wu, Yuki Asano, Pascal Mettes, Basura Fernando, Björn Ommer, Cees Snoek


Abstract
Diffusion models are a promising tool for high-quality text generation. However, current models face multiple drawbacks including slow sampling, noise schedule sensitivity, and misalignment between the training and sampling stages. In this paper, we introduce FlowSeq, which bypasses all current drawbacks by leveraging flow matching for conditional text generation. FlowSeq can generate text in a few steps by training with a novel anchor loss, alleviating the need for expensive hyperparameter optimization of the noise schedule prevalent in diffusion models. We extensively evaluate our proposed method and show competitive performance in tasks such as question generation, open-domain dialogue, and paraphrasing tasks.
Anthology ID:
2024.eacl-short.33
Volume:
Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2: Short Papers)
Month:
March
Year:
2024
Address:
St. Julian’s, Malta
Editors:
Yvette Graham, Matthew Purver
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
380–392
Language:
URL:
https://aclanthology.org/2024.eacl-short.33
DOI:
Bibkey:
Cite (ACL):
Vincent Hu, Di Wu, Yuki Asano, Pascal Mettes, Basura Fernando, Björn Ommer, and Cees Snoek. 2024. Flow Matching for Conditional Text Generation in a Few Sampling Steps. In Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2: Short Papers), pages 380–392, St. Julian’s, Malta. Association for Computational Linguistics.
Cite (Informal):
Flow Matching for Conditional Text Generation in a Few Sampling Steps (Hu et al., EACL 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.eacl-short.33.pdf
Software:
 2024.eacl-short.33.software.zip