Parameter-Efficient Abstractive Question Answering over Tables or Text

Open Access
Authors
Publication date 2022
Host editors
  • S. Feng
  • H. Wan
  • C. Yuan
  • H. Yu
Book title Proceedings of the Second DialDoc Workshop on Document-grounded Dialogue and Conversational Question Answering
Book subtitle proceedings of the workshop : DialDoc 2022 : May 26, 2022
ISBN (electronic)
  • 9781955917339
Event 2nd DialDoc Workshop on Document-grounded Dialogue and Conversational Question Answering
Pages (from-to) 41–53
Publisher Stroudsburg, PA: Association for Computational Linguistics
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
A long-term ambition of information seeking QA systems is to reason over multi-modal contexts and generate natural answers to user queries. Today, memory intensive pre-trained language models are adapted to downstream tasks such as QA by fine-tuning the model on QA data in a specific modality like unstructured text or structured tables. To avoid training such memory-hungry models while utilizing a uniform architecture for each modality, parameter-efficient adapters add and train small task-specific bottle-neck layers between transformer layers. In this work, we study parameter-efficient abstractive QA in encoder-decoder models over structured tabular data and unstructured textual data using only 1.5% additional parameters for each modality. We also ablate over adapter layers in both encoder and decoder modules to study the efficiency-performance trade-off and demonstrate that reducing additional trainable parameters down to 0.7%-1.0% leads to comparable results. Our models out-perform current state-of-the-art models on tabular QA datasets such as Tablesum and FeTaQA, and achieve comparable performance on a textual QA dataset such as NarrativeQA using significantly less trainable parameters than fine-tuning.
Document type Conference contribution
Note With supplementary video
Language English
Published at https://doi.org/10.18653/v1/2022.dialdoc-1.5
Other links https://paperswithcode.com/paper/parameter-efficient-abstractive-question-1
Downloads
2022.dialdoc-1.5 (Final published version)
Supplementary materials
Permalink to this page
Back