AUTH @ CLSciSumm 20, LaySumm 20, LongSumm 20

Alexios Gidiotis, Stefanos Stefanidis, Grigorios Tsoumakas


Abstract
We present the systems we submitted for the shared tasks of the Workshop on Scholarly Document Processing at EMNLP 2020. Our approaches to the tasks are focused on exploiting large Transformer models pre-trained on huge corpora and adapting them to the different shared tasks. For tasks 1A and 1B of CL-SciSumm we are using different variants of the BERT model to tackle the tasks of “cited text span” and “facet” identification. For the summarization tasks 2 of CL-SciSumm, LaySumm and LongSumm we make use of different variants of the PEGASUS model, with and without fine-tuning, adapted to the nuances of each one of those particular tasks.
Anthology ID:
2020.sdp-1.28
Volume:
Proceedings of the First Workshop on Scholarly Document Processing
Month:
November
Year:
2020
Address:
Online
Editors:
Muthu Kumar Chandrasekaran, Anita de Waard, Guy Feigenblat, Dayne Freitag, Tirthankar Ghosal, Eduard Hovy, Petr Knoth, David Konopnicki, Philipp Mayr, Robert M. Patton, Michal Shmueli-Scheuer
Venue:
sdp
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
251–260
Language:
URL:
https://aclanthology.org/2020.sdp-1.28
DOI:
10.18653/v1/2020.sdp-1.28
Bibkey:
Cite (ACL):
Alexios Gidiotis, Stefanos Stefanidis, and Grigorios Tsoumakas. 2020. AUTH @ CLSciSumm 20, LaySumm 20, LongSumm 20. In Proceedings of the First Workshop on Scholarly Document Processing, pages 251–260, Online. Association for Computational Linguistics.
Cite (Informal):
AUTH @ CLSciSumm 20, LaySumm 20, LongSumm 20 (Gidiotis et al., sdp 2020)
Copy Citation:
PDF:
https://aclanthology.org/2020.sdp-1.28.pdf
Video:
 https://slideslive.com/38941222
Data
ScisummNet