NonFactS: NonFactual Summary Generation for Factuality Evaluation in Document Summarization

Open Access
Authors
Publication date 2023
Host editors
  • A. Rogers
  • J. Boyd-Graber
  • N. Okazaki
Book title Findings of the Association for Computational Linguistics: ACL 2023
Book subtitle July 9-14, 2023
ISBN (electronic)
  • 9781959429623
Event 61st Annual Meeting of the Association for Computational Linguistics
Pages (from-to) 6405-6419
Publisher Stroudsburg, PA: Association for Computational Linguistics
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Pre-trained abstractive summarization models can generate fluent summaries and achieve high ROUGE scores. Previous research has found that these models often generate summaries that are inconsistent with their context document and contain nonfactual information. To evaluate factuality in document summarization, a document-level Natural Language Inference (NLI) classifier can be used. However, training such a classifier requires large-scale high-quality factual and nonfactual samples. To that end, we introduce NonFactS, a data generation model, to synthesize nonfactual summaries given a context document and a human-annotated (reference) factual summary. Compared to previous methods, our nonfactual samples are more abstractive and more similar to their corresponding factual samples, resulting in state-of-the-art performance on two factuality evaluation benchmarks, FALSESUM and SUMMAC. Our experiments demonstrate that even without human-annotated summaries, NonFactS can use random sentences to generate nonfactual summaries and a classifier trained on these samples generalizes to out-of-domain documents.
Document type Conference contribution
Language English
Published at https://doi.org/10.18653/v1/2023.findings-acl.400
Other links https://aclanthology.org/2023.findings-acl.400.mp4 https://github.com/asoleimanib/NonFactS
Downloads
2023.findings-acl.400 (Final published version)
Permalink to this page
Back