Deep Attentive Sentence Ordering Network

Baiyun Cui, Yingming Li, Ming Chen, Zhongfei Zhang


Abstract
In this paper, we propose a novel deep attentive sentence ordering network (referred as ATTOrderNet) which integrates self-attention mechanism with LSTMs in the encoding of input sentences. It enables us to capture global dependencies among sentences regardless of their input order and obtains a reliable representation of the sentence set. With this representation, a pointer network is exploited to generate an ordered sequence. The proposed model is evaluated on Sentence Ordering and Order Discrimination tasks. The extensive experimental results demonstrate its effectiveness and superiority to the state-of-the-art methods.
Anthology ID:
D18-1465
Volume:
Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing
Month:
October-November
Year:
2018
Address:
Brussels, Belgium
Editors:
Ellen Riloff, David Chiang, Julia Hockenmaier, Jun’ichi Tsujii
Venue:
EMNLP
SIG:
SIGDAT
Publisher:
Association for Computational Linguistics
Note:
Pages:
4340–4349
Language:
URL:
https://aclanthology.org/D18-1465
DOI:
10.18653/v1/D18-1465
Bibkey:
Cite (ACL):
Baiyun Cui, Yingming Li, Ming Chen, and Zhongfei Zhang. 2018. Deep Attentive Sentence Ordering Network. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 4340–4349, Brussels, Belgium. Association for Computational Linguistics.
Cite (Informal):
Deep Attentive Sentence Ordering Network (Cui et al., EMNLP 2018)
Copy Citation:
PDF:
https://aclanthology.org/D18-1465.pdf
Video:
 https://aclanthology.org/D18-1465.mp4