UMSE: Unified Multi-scenario Summarization Evaluation

Shen Gao, Zhitao Yao, Chongyang Tao, Xiuying Chen, Pengjie Ren, Zhaochun Ren, Zhumin Chen


Abstract
Summarization quality evaluation is a non-trivial task in text summarization. Contemporary methods can be mainly categorized into two scenarios: (1) reference-based: evaluating with human-labeled reference summary; (2) reference-free: evaluating the summary consistency of the document. Recent studies mainly focus on one of these scenarios and explore training neural models built on PLMs to align with human criteria. However, the models from different scenarios are optimized individually, which may result in sub-optimal performance since they neglect the shared knowledge across different scenarios. Besides, designing individual models for each scenario caused inconvenience to the user. Inspired by this, we propose Unified Multi-scenario Summarization Evaluation Model (UMSE). More specifically, we propose a perturbed prefix tuning method to share cross-scenario knowledge between scenarios and use a self-supervised training paradigm to optimize the model without extra human labeling. Our UMSE is the first unified summarization evaluation framework engaged with the ability to be used in three evaluation scenarios. Experimental results across three typical scenarios on the benchmark dataset SummEval indicate that our UMSE can achieve comparable performance with several existing strong methods which are specifically designed for each scenario.
Anthology ID:
2023.findings-acl.236
Volume:
Findings of the Association for Computational Linguistics: ACL 2023
Month:
July
Year:
2023
Address:
Toronto, Canada
Editors:
Anna Rogers, Jordan Boyd-Graber, Naoaki Okazaki
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
3837–3849
Language:
URL:
https://aclanthology.org/2023.findings-acl.236
DOI:
10.18653/v1/2023.findings-acl.236
Bibkey:
Cite (ACL):
Shen Gao, Zhitao Yao, Chongyang Tao, Xiuying Chen, Pengjie Ren, Zhaochun Ren, and Zhumin Chen. 2023. UMSE: Unified Multi-scenario Summarization Evaluation. In Findings of the Association for Computational Linguistics: ACL 2023, pages 3837–3849, Toronto, Canada. Association for Computational Linguistics.
Cite (Informal):
UMSE: Unified Multi-scenario Summarization Evaluation (Gao et al., Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-acl.236.pdf