On the Intractability to Synthesize Factual Inconsistencies in Summarization

Ge Luo, Weisi Fan, Miaoran Li, Youbiao He, Yinfei Yang, Forrest Bao


Abstract
Factual consistency detection has gotten raised attention in the task of abstractive summarization. Many existing works rely on synthetic training data, which may not accurately reflect or match the inconsistencies produced by summarization models. In this paper, we first systematically analyze the shortcomings of the current methods in synthesizing inconsistent summaries. Current synthesis methods may fail to produce inconsistencies of coreference errors and discourse errors, per our quantitative and qualitative study. Then, employing the parameter-efficient finetuning (PEFT) technique, we discover that a competitive factual consistency detector can be achieved using thousands of real model-generated summaries with human annotations. Our study demonstrates the importance of real machine-generated texts with human annotation in NLG evaluation as our model outperforms the SOTA on the CoGenSumm, FactCC, Frank, and SummEval datasets.
Anthology ID:
2024.findings-eacl.69
Volume:
Findings of the Association for Computational Linguistics: EACL 2024
Month:
March
Year:
2024
Address:
St. Julian’s, Malta
Editors:
Yvette Graham, Matthew Purver
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1026–1037
Language:
URL:
https://aclanthology.org/2024.findings-eacl.69
DOI:
Bibkey:
Cite (ACL):
Ge Luo, Weisi Fan, Miaoran Li, Youbiao He, Yinfei Yang, and Forrest Bao. 2024. On the Intractability to Synthesize Factual Inconsistencies in Summarization. In Findings of the Association for Computational Linguistics: EACL 2024, pages 1026–1037, St. Julian’s, Malta. Association for Computational Linguistics.
Cite (Informal):
On the Intractability to Synthesize Factual Inconsistencies in Summarization (Luo et al., Findings 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.findings-eacl.69.pdf