Semi-supervised Parsing with a Variational Autoencoding Parser

Xiao Zhang, Dan Goldwasser


Abstract
We propose an end-to-end variational autoencoding parsing (VAP) model for semi-supervised graph-based projective dependency parsing. It encodes the input using continuous latent variables in a sequential manner by deep neural networks (DNN) that can utilize the contextual information, and reconstruct the input using a generative model. The VAP model admits a unified structure with different loss functions for labeled and unlabeled data with shared parameters. We conducted experiments on the WSJ data sets, showing the proposed model can use the unlabeled data to increase the performance on a limited amount of labeled data, on a par with a recently proposed semi-supervised parser with faster inference.
Anthology ID:
2020.iwpt-1.5
Volume:
Proceedings of the 16th International Conference on Parsing Technologies and the IWPT 2020 Shared Task on Parsing into Enhanced Universal Dependencies
Month:
July
Year:
2020
Address:
Online
Editors:
Gosse Bouma, Yuji Matsumoto, Stephan Oepen, Kenji Sagae, Djamé Seddah, Weiwei Sun, Anders Søgaard, Reut Tsarfaty, Dan Zeman
Venue:
IWPT
SIG:
SIGPARSE
Publisher:
Association for Computational Linguistics
Note:
Pages:
40–47
Language:
URL:
https://aclanthology.org/2020.iwpt-1.5
DOI:
10.18653/v1/2020.iwpt-1.5
Bibkey:
Cite (ACL):
Xiao Zhang and Dan Goldwasser. 2020. Semi-supervised Parsing with a Variational Autoencoding Parser. In Proceedings of the 16th International Conference on Parsing Technologies and the IWPT 2020 Shared Task on Parsing into Enhanced Universal Dependencies, pages 40–47, Online. Association for Computational Linguistics.
Cite (Informal):
Semi-supervised Parsing with a Variational Autoencoding Parser (Zhang & Goldwasser, IWPT 2020)
Copy Citation:
PDF:
https://aclanthology.org/2020.iwpt-1.5.pdf
Video:
 http://slideslive.com/38929672