Exploiting Discourse-Level Segmentation for Extractive Summarization

Zhengyuan Liu, Nancy Chen


Abstract
Extractive summarization selects and concatenates the most essential text spans in a document. Most, if not all, neural approaches use sentences as the elementary unit to select content for summarization. However, semantic segments containing supplementary information or descriptive details are often nonessential in the generated summaries. In this work, we propose to exploit discourse-level segmentation as a finer-grained means to more precisely pinpoint the core content in a document. We investigate how the sub-sentential segmentation improves extractive summarization performance when content selection is modeled through two basic neural network architectures and a deep bi-directional transformer. Experiment results on the CNN/Daily Mail dataset show that discourse-level segmentation is effective in both cases. In particular, we achieve state-of-the-art performance when discourse-level segmentation is combined with our adapted contextual representation model.
Anthology ID:
D19-5415
Volume:
Proceedings of the 2nd Workshop on New Frontiers in Summarization
Month:
November
Year:
2019
Address:
Hong Kong, China
Venues:
EMNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
116–121
Language:
URL:
https://aclanthology.org/D19-5415
DOI:
10.18653/v1/D19-5415
Bibkey:
Cite (ACL):
Zhengyuan Liu and Nancy Chen. 2019. Exploiting Discourse-Level Segmentation for Extractive Summarization. In Proceedings of the 2nd Workshop on New Frontiers in Summarization, pages 116–121, Hong Kong, China. Association for Computational Linguistics.
Cite (Informal):
Exploiting Discourse-Level Segmentation for Extractive Summarization (Liu & Chen, EMNLP 2019)
Copy Citation:
PDF:
https://aclanthology.org/D19-5415.pdf