@InProceedings{han-jiang-tu:2017:EMNLP2017,
  author    = {Han, Wenjuan  and  Jiang, Yong  and  Tu, Kewei},
  title     = {Dependency Grammar Induction with Neural Lexicalization and Big Training Data},
  booktitle = {Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing},
  month     = {September},
  year      = {2017},
  address   = {Copenhagen, Denmark},
  publisher = {Association for Computational Linguistics},
  pages     = {1683--1688},
  abstract  = {We study the impact of big models (in terms of the degree of lexicalization)
	and big data (in terms of the training corpus size) on dependency grammar
	induction.
	We experimented with L-DMV, a lexicalized version of Dependency Model with
	Valence \cite{Klein:2004:CIS:1218955.1219016} and L-NDMV, our lexicalized
	extension of the Neural Dependency Model with Valence
	\cite{jiang-han-tu:2016:EMNLP2016}. 
	We find that L-DMV only benefits from very small degrees of lexicalization and
	moderate sizes of training corpora. L-NDMV can benefit from big training data
	and lexicalization of greater degrees, especially when enhanced with good model
	initialization, and it achieves a result that is competitive with the current
	state-of-the-art.},
  url       = {https://www.aclweb.org/anthology/D17-1176}
}

