Flow Matching for Conditional Text Generation in a Few Sampling Steps

Open Access
Authors
  • Fernando Fernández-Méndez
  • B. Ommer
  • C.G.M. Snoek ORCID logo
Publication date 2024
Host editors
  • Y. Graham
  • M. Purver
Book title The 18th Conference of the European Chapter of the Association for Computational Linguistics
Book subtitle proceedings of the conference : EACL 2024 : March 17-22, 2024
ISBN (electronic)
  • 9798891760899
Event 18th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2024
Volume | Issue number 2
Pages (from-to) 380-392
Publisher Kerrville, TX: Association for Computational Linguistics
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Diffusion models are a promising tool for high-quality text generation. However, current models face multiple drawbacks including slow sampling, noise schedule sensitivity, and misalignment between the training and sampling stages. In this paper, we introduce FlowSeq, which bypasses all current drawbacks by leveraging flow matching for conditional text generation. FlowSeq can generate text in a few steps by training with a novel anchor loss, alleviating the need for expensive hyperparameter optimization of the noise schedule prevalent in diffusion models. We extensively evaluate our proposed method and show competitive performance in tasks such as question generation, open-domain dialogue, and paraphrasing tasks.
Document type Conference contribution
Note With software
Language English
Published at https://doi.org/10.18653/v1/2024.eacl-short.33
Downloads
2024.eacl-short.33 (Final published version)
Supplementary materials
Permalink to this page
Back