Meta-Learning for Domain Generalization in Semantic Parsing

Open Access
Authors
Publication date 2021
Host editors
  • K. Toutanova
  • A. Rumshisky
  • L. Zettlemoyer
  • D. Hakkani-Tur
  • I. Beltagy
  • S. Bethard
  • R. Cotterell
  • T. Chakraborty
  • Y. Zhou
Book title The 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
Book subtitle NAACL-HLT 2021 : proceedings of the conference : June 6-11, 2021
ISBN (electronic)
  • 9781954085466
Event 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021
Pages (from-to) 366-379
Number of pages 14
Publisher Stroudsburg, PA: The Association for Computational Linguistics
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract

The importance of building semantic parsers which can be applied to new domains and generate programs unseen at training has long been acknowledged, and datasets testing out-of-domain performance are becoming increasingly available. However, little or no attention has been devoted to learning algorithms or objectives which promote domain generalization, with virtually all existing approaches relying on standard supervised learning. In this work, we use a meta-learning framework which targets zero-shot domain generalization for semantic parsing. We apply a model-agnostic training algorithm that simulates zero-shot parsing by constructing virtual train and test sets from disjoint domains. The learning objective capitalizes on the intuition that gradient steps that improve source-domain performance should also improve target-domain performance, thus encouraging a parser to generalize to unseen target domains. Experimental results on the (English) Spider and Chinese Spider datasets show that the meta-learning objective significantly boosts the performance of a baseline parser.

Document type Conference contribution
Note With supplementary video
Language English
Published at https://doi.org/10.18653/v1/2021.naacl-main.33
Other links https://www.scopus.com/pages/publications/85111187772
Downloads
2021.naacl-main.33 (Final published version)
Supplementary materials
Permalink to this page
Back