Abstract

In the summarization domain, a key requirement for summaries is to be factually consistent with the input document. Previous work has found that natural language inference (NLI) models do not perform competitively when applied to inconsistency detection. In this work, we revisit the use of NLI for inconsistency detection, finding that past work suffered from a mismatch in input granularity between NLI datasets (sentence-level), and inconsistency detection (document level). We provide a highly effective and light-weight method called SummaCConv that enables NLI models to be successfully used for this task by segmenting documents into sentence units and aggregating scores between pairs of sentences. We furthermore introduce a new benchmark called SummaC (Summary Consistency) which consists of six large inconsistency detection datasets. On this dataset, SummaCConv obtains state-of-the-art results with a balanced accuracy of 74.4%, a 5% improvement compared with prior work.

Details

Title
SummaC: Re-Visiting NLI-based Models for Inconsistency Detection in Summarization
Author
Laban, Philippe; Schnabel, Tobias; Bennett, Paul N; Hearst, Marti A
Pages
163-177
Publication year
2022
Publication date
2022
Publisher
MIT Press Journals, The
ISSN
2307387X
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
2893941543
Copyright
© 2022. This work is published under https://creativecommons.org/licenses/by/4.0/legalcode (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.