Table-To-Text generation and pre-training with TabT5

Ewa Andrejczuk, Julian Eisenschlos, Francesco Piccinno, Syrine Krichene, Yasemin Altun


Abstract
Encoder-only transformer models have been successfully applied to different table understanding tasks, as in TAPAS. A major limitation of these architectures is that they are constrained to classification-like tasks such as cell selection or entailment detection. We present TabT5, an encoder-decoder model that generates natural language text based on tables and textual inputs. TabT5 overcomes the encoder-only limitation by incorporating a decoder component and leverages the input structure with table specific embeddings and pre-training. TabT5 achieves new state-of-the-art results on several domains, including spreadsheet formula prediction with a 15% increase in sequence accuracy, QA with a 2.5% increase in sequence accuracy and data-to-text generation with a 2.5% increase in BLEU.
Anthology ID:
2022.findings-emnlp.503
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2022
Month:
December
Year:
2022
Address:
Abu Dhabi, United Arab Emirates
Editors:
Yoav Goldberg, Zornitsa Kozareva, Yue Zhang
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
6758–6766
Language:
URL:
https://aclanthology.org/2022.findings-emnlp.503
DOI:
10.18653/v1/2022.findings-emnlp.503
Bibkey:
Cite (ACL):
Ewa Andrejczuk, Julian Eisenschlos, Francesco Piccinno, Syrine Krichene, and Yasemin Altun. 2022. Table-To-Text generation and pre-training with TabT5. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 6758–6766, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
Cite (Informal):
Table-To-Text generation and pre-training with TabT5 (Andrejczuk et al., Findings 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.findings-emnlp.503.pdf
Video:
 https://aclanthology.org/2022.findings-emnlp.503.mp4