<?xml version="1.0" encoding="UTF-8" ?>
<volume id="W17">
  <paper id="1000">
    <title>Proceedings of the MultiLing 2017 Workshop on Summarization and Summary Evaluation Across Source Types and Genres</title>
    <editor>George Giannakopoulos</editor>
    <editor>Elena Lloret</editor>
    <editor>John M. Conroy</editor>
    <editor>Josef Steinberger</editor>
    <editor>Marina Litvak</editor>
    <editor>Peter Rankel</editor>
    <editor>Benoit Favre</editor>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <url>http://www.aclweb.org/anthology/W17-10</url>
    <bibtype>book</bibtype>
    <bibkey>MultiLing2017:2017</bibkey>
  </paper>

  <paper id="1001">
    <title>MultiLing 2017 Overview</title>
    <author><first>George</first><last>Giannakopoulos</last></author>
    <author><first>John</first><last>Conroy</last></author>
    <author><first>Jeff</first><last>Kubina</last></author>
    <author><first>Peter A.</first><last>Rankel</last></author>
    <author><first>Elena</first><last>Lloret</last></author>
    <author><first>Josef</first><last>Steinberger</last></author>
    <author><first>Marina</first><last>Litvak</last></author>
    <author><first>Benoit</first><last>Favre</last></author>
    <booktitle>Proceedings of the MultiLing 2017 Workshop on Summarization and Summary Evaluation Across Source Types and Genres</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>1&#8211;6</pages>
    <url>http://www.aclweb.org/anthology/W17-1001</url>
    <abstract>In this brief report we present an overview of the MultiLing 2017 effort and
	workshop, as implemented within EACL 2017.
	MultiLing is a community-driven initiative that pushes the state-of-the-art in
	Automatic Summarization by providing data sets and fostering further research
	and development of summarization systems.
	This year the scope of the workshop was widened, bringing together researchers
	that work on summarization across sources, languages and genres. We summarize
	the main tasks planned and implemented this year, the contributions received,
	and we also provide insights on next steps.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>giannakopoulos-EtAl:2017:MultiLing2017</bibkey>
  </paper>

  <paper id="1002">
    <title>Decoupling Encoder and Decoder Networks for Abstractive Document Summarization</title>
    <author><first>Ying</first><last>Xu</last></author>
    <author><first>Jey Han</first><last>Lau</last></author>
    <author><first>Timothy</first><last>Baldwin</last></author>
    <author><first>Trevor</first><last>Cohn</last></author>
    <booktitle>Proceedings of the MultiLing 2017 Workshop on Summarization and Summary Evaluation Across Source Types and Genres</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>7&#8211;11</pages>
    <url>http://www.aclweb.org/anthology/W17-1002</url>
    <abstract>Abstractive document summarization seeks to automatically generate a summary
	for a document, based on some abstract &#x201d;understanding&#x201d; of the original
	document. State-of-the-art techniques traditionally use
	attentive encoder&#8211;decoder architectures.  However, due to the large number of
	parameters in these models, they require large training datasets and long
	training times. In this paper, we propose decoupling the encoder and decoder
	networks, and training them separately.  We encode documents using an
	unsupervised document encoder, and then feed the document vector to a recurrent
	neural network decoder. With this decoupled architecture, we decrease the
	number of parameters in the decoder substantially, and shorten its training
	time.  Experiments show that the decoupled model achieves comparable
	performance with state-of-the-art models for in-domain documents, but less well
	for out-of-domain documents.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>xu-EtAl:2017:MultiLing2017</bibkey>
  </paper>

  <paper id="1003">
    <title>Centroid-based Text Summarization through Compositionality of Word Embeddings</title>
    <author><first>Gaetano</first><last>Rossiello</last></author>
    <author><first>Pierpaolo</first><last>Basile</last></author>
    <author><first>Giovanni</first><last>Semeraro</last></author>
    <booktitle>Proceedings of the MultiLing 2017 Workshop on Summarization and Summary Evaluation Across Source Types and Genres</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>12&#8211;21</pages>
    <url>http://www.aclweb.org/anthology/W17-1003</url>
    <abstract>The textual similarity is a crucial aspect for many extractive text
	summarization methods. A bag-of-words representation does not allow to grasp
	the semantic relationships between concepts when comparing strongly related
	sentences with no words in common. To overcome this issue, in this paper we
	propose a centroid-based method for text summarization that exploits the
	compositional capabilities of word embeddings. The evaluations on
	multi-document and multilingual datasets prove the effectiveness of the
	continuous vector representation of words compared to the bag-of-words model.
	Despite its simplicity, our method achieves good performance even in comparison
	to more complex deep learning models. Our method is unsupervised and it can be
	adopted in other summarization tasks.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>rossiello-basile-semeraro:2017:MultiLing2017</bibkey>
  </paper>

  <paper id="1004">
    <title>Query-based summarization using MDL principle</title>
    <author><first>Marina</first><last>Litvak</last></author>
    <author><first>Natalia</first><last>Vanetik</last></author>
    <booktitle>Proceedings of the MultiLing 2017 Workshop on Summarization and Summary Evaluation Across Source Types and Genres</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>22&#8211;31</pages>
    <url>http://www.aclweb.org/anthology/W17-1004</url>
    <abstract>Query-based text summarization is aimed at extracting essential information
	that answers the query from original text. The answer is presented  
	in a minimal, often predefined, number of words. In this paper we introduce a
	new unsupervised approach for query-based extractive summarization, based on
	the minimum description length (MDL) principle that employs Krimp compression
	algorithm (Vreeken et al., 2011). The key idea of our approach is to select
	frequent word sets related to a given query that compress document sentences
	better and therefore describe the document better.
	A summary is extracted by selecting sentences that best cover query-related
	frequent word sets.
	The approach is evaluated based on the DUC 2005 and DUC 2006 datasets which are
	specifically designed for query-based summarization (DUC, 2005 2006). It
	competes with the best results.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>litvak-vanetik:2017:MultiLing2017</bibkey>
  </paper>

  <paper id="1005">
    <title>Word Embedding and Topic Modeling Enhanced Multiple Features for Content Linking and Argument / Sentiment Labeling in Online Forums</title>
    <author><first>Lei</first><last>Li</last></author>
    <author><first>Liyuan</first><last>Mao</last></author>
    <author><first>Moye</first><last>Chen</last></author>
    <booktitle>Proceedings of the MultiLing 2017 Workshop on Summarization and Summary Evaluation Across Source Types and Genres</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>32&#8211;36</pages>
    <url>http://www.aclweb.org/anthology/W17-1005</url>
    <abstract>Multiple grammatical and semantic features are adopted in content linking and
	argument/sentiment labeling for online forums in this paper. There are mainly
	two different methods for content linking. First, we utilize the deep feature
	obtained from Word Embedding Model in deep learning and compute sentence
	similarity. Second, we use multiple traditional features to locate candidate
	linking sentences, and then adopt a voting method to obtain the final result.
	LDA topic modeling is used to mine latent semantic feature and K-means
	clustering is implemented for argument labeling, while features from sentiment
	dictionaries and rule-based sentiment analysis are integrated for sentiment
	labeling. Experimental results have shown that our methods are valid.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>li-mao-chen:2017:MultiLing2017</bibkey>
  </paper>

  <paper id="1006">
    <title>Ultra-Concise Multi-genre Summarisation of Web2.0: towards Intelligent Content Generation</title>
    <author><first>Elena</first><last>Lloret</last></author>
    <author><first>Ester</first><last>Boldrini</last></author>
    <author><first>Patricio</first><last>Martinez-Barco</last></author>
    <author><first>Manuel</first><last>Palomar</last></author>
    <booktitle>Proceedings of the MultiLing 2017 Workshop on Summarization and Summary Evaluation Across Source Types and Genres</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>37&#8211;46</pages>
    <url>http://www.aclweb.org/anthology/W17-1006</url>
    <abstract>The electronic Word of Mouth has become the most powerful communication channel
	thanks to the wide usage of the Social Media. Our research proposes an approach
	towards the production of automatic ultra-concise summaries from multiple Web
	2.0
	sources. We exploit user-generated content from reviews and microblogs in dif-
	ferent domains, and compile and analyse four types of ultra-concise summaries:
	a)positive information, b) negative information; c) both or d) objective
	information. The appropriateness and usefulness of our model is demonstrated by
	its successful results and great potential in real-life applications, thus
	meaning a relevant advancement of the state-of-the-art approaches.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>lloret-EtAl:2017:MultiLing2017</bibkey>
  </paper>

  <paper id="1007">
    <title>Machine Learning Approach to Evaluate MultiLingual Summaries</title>
    <author><first>Samira</first><last>Ellouze</last></author>
    <author><first>Maher</first><last>Jaoua</last></author>
    <author><first>Lamia</first><last>Hadrich Belguith</last></author>
    <booktitle>Proceedings of the MultiLing 2017 Workshop on Summarization and Summary Evaluation Across Source Types and Genres</booktitle>
    <month>April</month>
    <year>2017</year>
    <address>Valencia, Spain</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>47&#8211;54</pages>
    <url>http://www.aclweb.org/anthology/W17-1007</url>
    <abstract>The present paper introduces a new MultiLing text summary evaluation method.
	This method relies on machine learning approach which operates by combining
	multiple features to build models that predict the human score (overall
	responsiveness) of a new summary. We have tried several single and &#x201c;ensemble
	learning&#x201d; classifiers to build the best model. We have experimented our
	method
	in summary level evaluation where we evaluate each text summary separately. The
	correlation between built models and human score is better than the correlation
	between baselines and manual score.</abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>ellouze-jaoua-hadrichbelguith:2017:MultiLing2017</bibkey>
  </paper>

</volume>

