<?xml version='1.0' encoding='UTF-8'?> 
<volume id='S17'>
  <paper id='1000'>
    <title>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </title>
    <editor>Nancy Ide</editor>
    <editor>Aurélie Herbelot</editor>
    <editor>Lluís Màrquez</editor>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <url>http://www.aclweb.org/anthology/S17-1</url>
    <doi>10.18653/v1/S17-1</doi>
    <bibtype>book</bibtype>
    <bibkey>starSEM:2017</bibkey>
  </paper>
  <paper id='1001'>
    <title>What Analogies Reveal about Word Vectors and their Compositionality</title>
    <author>
      <first>Gregory</first>
      <last>Finley</last>
    </author>
    <author>
      <first>Stephanie</first>
      <last>Farmer</last>
    </author>
    <author>
      <first>Serguei</first>
      <last>Pakhomov</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>1–11</pages>
    <url>http://www.aclweb.org/anthology/S17-1001</url>
    <doi>10.18653/v1/S17-1001</doi>
    <abstract>
      Analogy completion via vector arithmetic has become a common means of
      demonstrating the compositionality of word embeddings. Previous work have
      shown that this strategy works more reliably for certain types of
      analogical word relationships than for others, but these studies have not
      offered a convincing account for why this is the case. We arrive at such
      an account through an experiment that targets a wide variety of analogy
      questions and defines a baseline condition to more accurately measure the
      efficacy of our system. We find that the most reliably solvable analogy
      categories involve either 1) the application of a morpheme with clear
      syntactic effects, 2) male–female alternations, or 3) named entities.
      These broader types do not pattern cleanly along a syntactic–semantic
      divide. We suggest instead that their commonality is distributional, in
      that the difference between the distributions of two words in any given
      pair encompasses a relatively small number of word types. Our study offers
      a needed explanation for why analogy tests succeed and fail where they do
      and provides nuanced insight into the relationship between word
      distributions and the theoretical linguistic domains of syntax and
      semantics. 
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>finley-farmer-pakhomov:2017:starSEM</bibkey>
  </paper>
  <paper id='1002'>
    <title>Learning Antonyms with Paraphrases and a Morphology-Aware Neural Network</title>
    <author>
      <first>Sneha</first>
      <last>Rajana</last>
    </author>
    <author>
      <first>Chris</first>
      <last>Callison-Burch</last>
    </author>
    <author>
      <first>Marianna</first>
      <last>Apidianaki</last>
    </author>
    <author>
      <first>Vered</first>
      <last>Shwartz</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>12–21</pages>
    <url>http://www.aclweb.org/anthology/S17-1002</url>
    <doi>10.18653/v1/S17-1002</doi>
    <abstract>
      Recognizing and distinguishing antonyms from other types of semantic
      relations is an essential part of language understanding systems. In this
      paper, we present a novel method for deriving antonym pairs using
      paraphrase pairs containing negation markers. We further propose a neural
      network model, AntNET, that integrates morphological features indicative
      of antonymy into a path-based relation detection algorithm. We demonstrate
      that our model outperforms state-of-the-art models in distinguishing
      antonyms from other semantic relations and is capable of efficiently
      handling multi-word expressions.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>rajana-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1003'>
    <title>Decoding Sentiment from Distributed Representations of Sentences</title>
    <author>
      <first>Edoardo Maria</first>
      <last>Ponti</last>
    </author>
    <author>
      <first>Ivan</first>
      <last>Vulić</last>
    </author>
    <author>
      <first>Anna</first>
      <last>Korhonen</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>22–32</pages>
    <url>http://www.aclweb.org/anthology/S17-1003</url>
    <doi>10.18653/v1/S17-1003</doi>
    <abstract>
      Distributed representations of sentences have been developed recently to
      represent their meaning as real-valued vectors. However, it is not clear
      how much information such representations retain about the polarity of
      sentences. To study this question, we decode sentiment from unsupervised
      sentence representations learned with different architectures (sensitive
      to the order of words, the order of sentences, or none) in 9 typologically
      diverse languages. Sentiment results from the (recursive) composition of
      lexical items and grammatical strategies such as negation and concession.
      The results are manifold: we show that there is no `one-size-fits-all'
      representation architecture outperforming the others across the board.
      Rather, the top-ranking architectures depend on the language at hand.
      Moreover, we find that in several cases the additive composition model
      based on skip-gram word vectors may surpass supervised state-of-art
      architectures such as bi-directional LSTMs. Finally, we provide a possible
      explanation of the observed variation based on the type of negative
      constructions in each language.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>ponti-vulic-korhonen:2017:starSEM</bibkey>
  </paper>
  <paper id='1004'>
    <title>
      Detecting Asymmetric Semantic Relations in Context: A Case-Study on
      Hypernymy Detection
    </title>
    <author>
      <first>Yogarshi</first>
      <last>Vyas</last>
    </author>
    <author>
      <first>Marine</first>
      <last>Carpuat</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>33–43</pages>
    <url>http://www.aclweb.org/anthology/S17-1004</url>
    <doi>10.18653/v1/S17-1004</doi>
    <abstract>
      We introduce WHiC, a challenging testbed for detecting hypernymy, an
      asymmetric relation between words. While previous work has focused on
      detecting hypernymy between word types, we ground the meaning of words in
      specific contexts drawn from WordNet examples, and require predictions to
      be sensitive to changes in contexts. WHiC lets us analyze complementary
      properties of two approaches of inducing vector representations of word
      meaning in context. We show that such contextualized word representations
      also improve detection of a wider range of semantic relations in context.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>vyas-carpuat:2017:starSEM</bibkey>
  </paper>
  <paper id='1005'>
    <title>Domain-Specific New Words Detection in Chinese</title>
    <author>
      <first>Ao</first>
      <last>Chen</last>
    </author>
    <author>
      <first>Maosong</first>
      <last>Sun</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>44–53</pages>
    <url>http://www.aclweb.org/anthology/S17-1005</url>
    <doi>10.18653/v1/S17-1005</doi>
    <abstract>
      With the explosive growth of Internet, more and more domain-specific
      environments appear, such as forums, blogs, MOOCs and etc. Domain-specific
      words appear in these areas and always play a critical role in the
      domain-specific NLP tasks. This paper aims at extracting Chinese
      domain-specific new words automatically. The extraction of domain-specific
      new words has two parts including both new words in this domain and the
      especially important words. In this work, we propose a joint statistical
      model to perform these two works simultaneously. Compared to traditional
      new words detection models, our model doesn't need handcraft features
      which are labor intensive. Experimental results demonstrate that our joint
      model achieves a better performance compared with the state-of-the-art
      methods.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>chen-sun:2017:starSEM</bibkey>
  </paper>
  <paper id='1006'>
    <title>Deep Learning Models For Multiword Expression Identification</title>
    <author>
      <first>Waseem</first>
      <last>Gharbieh</last>
    </author>
    <author>
      <first>Virendrakumar</first>
      <last>Bhavsar</last>
    </author>
    <author>
      <first>Paul</first>
      <last>Cook</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>54–64</pages>
    <url>http://www.aclweb.org/anthology/S17-1006</url>
    <doi>10.18653/v1/S17-1006</doi>
    <abstract>
      Multiword expressions (MWEs) are lexical items that can be decomposed into
      multiple component words, but have properties that are unpredictable with
      respect to their component words. In this paper we propose the first deep
      learning models for token-level identification of MWEs. Specifically, we
      consider a layered feedforward network, a recurrent neural network, and
      convolutional neural networks. In experimental results we show that
      convolutional neural networks are able to outperform the previous
      state-of-the-art for MWE identification, with a convolutional neural
      network with three hidden layers giving the best performance.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>gharbieh-bhavsar-cook:2017:starSEM</bibkey>
  </paper>
  <paper id='1007'>
    <title>Emotion Intensities in Tweets</title>
    <author>
      <first>Saif</first>
      <last>Mohammad</last>
    </author>
    <author>
      <first>Felipe</first>
      <last>Bravo-Marquez</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>65–77</pages>
    <url>http://www.aclweb.org/anthology/S17-1007</url>
    <doi>10.18653/v1/S17-1007</doi>
    <abstract>
      This paper examines the task of detecting intensity of emotion from text.
      We create the first datasets of tweets annotated for anger, fear, joy, and
      sadness intensities. We use a technique called best–worst scaling (BWS)
      that improves annotation consistency and obtains reliable fine-grained
      scores. We show that emotion-word hashtags often impact emotion intensity,
      usually conveying a more intense emotion. Finally, we create a benchmark
      regression system and conduct experiments to determine: which features are
      useful for detecting emotion intensity; and, the extent to which two
      emotions are similar in terms of how they manifest in language.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>mohammad-bravomarquez:2017:starSEM</bibkey>
  </paper>
  <paper id='1008'>
    <title>Deep Active Learning for Dialogue Generation</title>
    <author>
      <first>Nabiha</first>
      <last>Asghar</last>
    </author>
    <author>
      <first>Pascal</first>
      <last>Poupart</last>
    </author>
    <author>
      <first>Xin</first>
      <last>Jiang</last>
    </author>
    <author>
      <first>Hang</first>
      <last>Li</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>78–83</pages>
    <url>http://www.aclweb.org/anthology/S17-1008</url>
    <doi>10.18653/v1/S17-1008</doi>
    <abstract>
      We propose an online, end-to-end, neural generative conversational model
      for open-domain dialogue. It is trained using a unique combination of
      offline two-phase supervised learning and online human-in-the-loop active
      learning. While most existing research proposes offline supervision or
      hand-crafted reward functions for online reinforcement, we devise a novel
      interactive learning mechanism based on hamming-diverse beam search for
      response generation and one-character user-feedback at each step.
      Experiments show that our model inherently promotes the generation of
      semantically relevant and interesting responses, and can be used to train
      agents with customized personas, moods and conversational styles.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>asghar-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1009'>
    <title>Mapping the Paraphrase Database to WordNet</title>
    <author>
      <first>Anne</first>
      <last>Cocos</last>
    </author>
    <author>
      <first>Marianna</first>
      <last>Apidianaki</last>
    </author>
    <author>
      <first>Chris</first>
      <last>Callison-Burch</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>84–90</pages>
    <url>http://www.aclweb.org/anthology/S17-1009</url>
    <doi>10.18653/v1/S17-1009</doi>
    <abstract>
      WordNet has facilitated important research in natural language processing
      but its usefulness is somewhat limited by its relatively small lexical
      coverage. The Paraphrase Database (PPDB) covers 650 times more words, but
      lacks the semantic structure of WordNet that would make it more directly
      useful for downstream tasks. We present a method for mapping words from
      PPDB to WordNet synsets with 89% accuracy. The mapping also lays important
      groundwork for incorporating WordNet's relations into PPDB so as to
      increase its utility for semantic reasoning in applications.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>cocos-apidianaki-callisonburch:2017:starSEM</bibkey>
  </paper>
  <paper id='1010'>
    <title>Semantic Frame Labeling with Target-based Neural Model</title>
    <author>
      <first>Yukun</first>
      <last>Feng</last>
    </author>
    <author>
      <first>Dong</first>
      <last>Yu</last>
    </author>
    <author>
      <first>Jian</first>
      <last>Xu</last>
    </author>
    <author>
      <first>Chunhua</first>
      <last>Liu</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>91–96</pages>
    <url>http://www.aclweb.org/anthology/S17-1010</url>
    <doi>10.18653/v1/S17-1010</doi>
    <abstract>
      This paper explores the automatic learning of distributed representations
      of the target's context for semantic frame labeling with target-based
      neural model. We constrain the whole sentence as the model's input without
      feature extraction from the sentence. This is different from many previous
      works in which local feature extraction of the targets is widely used.
      This constraint makes the task harder, especially with long sentences, but
      also makes our model easily applicable to a range of resources and other
      similar tasks. We evaluate our model on several resources and get the
      state-of-the-art result on subtask 2 of SemEval 2015 task 15. Finally, we
      extend the task to word-sense disambiguation task and we also achieve a
      strong result in comparison to state-of-the-art work.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>feng-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1011'>
    <title>
      Frame-Based Continuous Lexical Semantics through Exponential Family Tensor
      Factorization and Semantic Proto-Roles
    </title>
    <author>
      <first>Francis</first>
      <last>Ferraro</last>
    </author>
    <author>
      <first>Adam</first>
      <last>Poliak</last>
    </author>
    <author>
      <first>Ryan</first>
      <last>Cotterell</last>
    </author>
    <author>
      <first>Benjamin</first>
      <last>Van Durme</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>97–103</pages>
    <url>http://www.aclweb.org/anthology/S17-1011</url>
    <doi>10.18653/v1/S17-1011</doi>
    <abstract>
      We study how different frame annotations complement one another when
      learning continuous lexical semantics. We learn the representations from a
      tensorized skip-gram model that consistently encodes syntactic-semantic
      content better, with multiple 10% gains over baselines.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>ferraro-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1012'>
    <title>
      Distributed Prediction of Relations for Entities: The Easy, The Difficult,
      and The Impossible
    </title>
    <author>
      <first>Abhijeet</first>
      <last>Gupta</last>
    </author>
    <author>
      <first>Gemma</first>
      <last>Boleda</last>
    </author>
    <author>
      <first>Sebastian</first>
      <last>Padó</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>104–109</pages>
    <url>http://www.aclweb.org/anthology/S17-1012</url>
    <doi>10.18653/v1/S17-1012</doi>
    <abstract>
      Word embeddings are supposed to provide easy access to semantic relations
      such as "male of" (man–woman). While this claim has been investigated for
      concepts, little is known about the distributional behavior of relations
      of (Named) Entities. We describe two word embedding-based models that
      predict values for relational attributes of entities, and analyse them.
      The task is challenging, with major performance differences between
      relations. Contrary to many NLP tasks, high difficulty for a relation does
      not result from low frequency, but from (a) one-to-many mappings; and (b)
      lack of context patterns expressing the relation that are easy to pick up
      by word embeddings.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>gupta-boleda-pado:2017:starSEM</bibkey>
  </paper>
  <paper id='1013'>
    <title>Comparing Approaches for Automatic Question Identification</title>
    <author>
      <first>Angel</first>
      <last>Maredia</last>
    </author>
    <author>
      <first>Kara</first>
      <last>Schechtman</last>
    </author>
    <author>
      <first>Sarah Ita</first>
      <last>Levitan</last>
    </author>
    <author>
      <first>Julia</first>
      <last>Hirschberg</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>110–114</pages>
    <url>http://www.aclweb.org/anthology/S17-1013</url>
    <doi>10.18653/v1/S17-1013</doi>
    <abstract>
      Collecting spontaneous speech corpora that are open-ended, yet topically
      constrained, is increasingly popular for research in spoken dialogue
      systems and speaker state, inter alia. Typically, these corpora are
      labeled by human annotators, either in the lab or through crowd-sourcing;
      however, this is cumbersome and time-consuming for large corpora. We
      present four different approaches to automatically tagging a corpus when
      general topics of the conversations are known. We develop these approaches
      on the Columbia X-Cultural Deception corpus and find accuracy that
      significantly exceeds the baseline. Finally, we conduct a cross-corpus
      evaluation by testing the best performing approach on the
      Columbia/SRI/Colorado corpus.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>maredia-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1014'>
    <title>
      Does Free Word Order Hurt? Assessing the Practical Lexical Function Model
      for Croatian
    </title>
    <author>
      <first>Zoran</first>
      <last>Medić</last>
    </author>
    <author>
      <first>Jan</first>
      <last>Šnajder</last>
    </author>
    <author>
      <first>Sebastian</first>
      <last>Padó</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>115–120</pages>
    <url>http://www.aclweb.org/anthology/S17-1014</url>
    <attachment type="poster">S17-1014.Poster.pdf</attachment>
    <doi>10.18653/v1/S17-1014</doi>
    <abstract>
      The Practical Lexical Function (PLF) model is a model of computational
      distributional semantics that attempts to strike a balance between
      expressivity and learnability in predicting phrase meaning and shows
      competitive results. We investigate how well the PLF carries over to free
      word order languages, given that it builds on observations of
      predicate-argument combinations that are harder to recover in free word
      order languages. We evaluate variants of the PLF for Croatian, using a new
      lexical substitution dataset. We find that the PLF works about as well for
      Croatian as for English, but demonstrate that its strength lies in
      modeling verbs, and that the free word order affects the less robust PLF
      variant.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>medic-vsnajder-pado:2017:starSEM</bibkey>
  </paper>
  <paper id='1015'>
    <title>A Mixture Model for Learning Multi-Sense Word Embeddings</title>
    <author>
      <first>Dai Quoc</first>
      <last>Nguyen</last>
    </author>
    <author>
      <first>Dat Quoc</first>
      <last>Nguyen</last>
    </author>
    <author>
      <first>Ashutosh</first>
      <last>Modi</last>
    </author>
    <author>
      <first>Stefan</first>
      <last>Thater</last>
    </author>
    <author>
      <first>Manfred</first>
      <last>Pinkal</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>121–127</pages>
    <url>http://www.aclweb.org/anthology/S17-1015</url>
    <doi>10.18653/v1/S17-1015</doi>
    <abstract>
      Word embeddings are now a standard technique for inducing meaning
      representations for words. For getting good representations, it is
      important to take into account different senses of a word. In this paper,
      we propose a mixture model for learning multi-sense word embeddings. Our
      model generalizes the previous works in that it allows to induce different
      weights of different senses of a word. The experimental results show that
      our model outperforms previous models on standard evaluation tasks.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>nguyen-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1016'>
    <title>Aligning Script Events with Narrative Texts</title>
    <author>
      <first>Simon</first>
      <last>Ostermann</last>
    </author>
    <author>
      <first>Michael</first>
      <last>Roth</last>
    </author>
    <author>
      <first>Stefan</first>
      <last>Thater</last>
    </author>
    <author>
      <first>Manfred</first>
      <last>Pinkal</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>128–134</pages>
    <url>http://www.aclweb.org/anthology/S17-1016</url>
    <doi>10.18653/v1/S17-1016</doi>
    <abstract>
      Script knowledge plays a central role in text understanding and is
      relevant for a variety of downstream tasks. In this paper, we consider two
      recent datasets which provide a rich and general representation of script
      events in terms of paraphrase sets. We introduce the task of mapping event
      mentions in narrative texts to such script event types, and present a
      model for this task that exploits rich linguistic representations as well
      as information on temporal ordering. The results of our experiments
      demonstrate that this complex task is indeed feasible.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>ostermann-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1017'>
    <title>The (too Many) Problems of Analogical Reasoning with Word Vectors</title>
    <author>
      <first>Anna</first>
      <last>Rogers</last>
    </author>
    <author>
      <first>Aleksandr</first>
      <last>Drozd</last>
    </author>
    <author>
      <first>Bofang</first>
      <last>Li</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>135–148</pages>
    <url>http://www.aclweb.org/anthology/S17-1017</url>
    <doi>10.18653/v1/S17-1017</doi>
    <abstract>
      This paper explores the possibilities of analogical reasoning with vector
      space models. Given two pairs of words with the same relation (e.g.
      man:woman :: king:queen), it was proposed that the offset between one pair
      of the corresponding word vectors can be used to identify the unknown
      member of the other pair (king - man + woman = queen). We argue against
      such "linguistic regularities" as a model for linguistic relations in
      vector space models and as a benchmark, and we show that the vector offset
      (as well as two other, better-performing methods) suffers from dependence
      on vector similarity.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>rogers-drozd-li:2017:starSEM</bibkey>
  </paper>
  <paper id='1018'>
    <title>
      Semantic Frames and Visual Scenes: Learning Semantic Role Inventories from
      Image and Video Descriptions
    </title>
    <author>
      <first>Ekaterina</first>
      <last>Shutova</last>
    </author>
    <author>
      <first>Andreas</first>
      <last>Wundsam</last>
    </author>
    <author>
      <first>Helen</first>
      <last>Yannakoudakis</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>149–154</pages>
    <url>http://www.aclweb.org/anthology/S17-1018</url>
    <doi>10.18653/v1/S17-1018</doi>
    <abstract>
      Frame-semantic parsing and semantic role labelling, that aim to
      automatically assign semantic roles to arguments of verbs in a sentence,
      have become an active strand of research in NLP. However, to date these
      methods have relied on a predefined inventory of semantic roles. In this
      paper, we present a method to automatically learn argument role
      inventories for verbs from large corpora of text, images and videos. We
      evaluate the method against manually constructed role inventories in
      FrameNet and show that the visual model outperforms the language-only
      model and operates with a high precision.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>shutova-wundsam-yannakoudakis:2017:starSEM</bibkey>
  </paper>
  <paper id='1019'>
    <title>Acquiring Predicate Paraphrases from News Tweets</title>
    <author>
      <first>Vered</first>
      <last>Shwartz</last>
    </author>
    <author>
      <first>Gabriel</first>
      <last>Stanovsky</last>
    </author>
    <author>
      <first>Ido</first>
      <last>Dagan</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>155–160</pages>
    <url>http://www.aclweb.org/anthology/S17-1019</url>
    <doi>10.18653/v1/S17-1019</doi>
    <abstract>
      We present a simple method for ever-growing extraction of predicate
      paraphrases from news headlines in Twitter. Analysis of the output of ten
      weeks of collection shows that the accuracy of paraphrases with different
      support levels is estimated between 60-86%. We also demonstrate that our
      resource is to a large extent complementary to existing resources,
      providing many novel paraphrases. Our resource is publicly available,
      continuously expanding based on daily news.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>shwartz-stanovsky-dagan:2017:starSEM</bibkey>
  </paper>
  <paper id='1020'>
    <title>Evaluating Semantic Parsing against a Simple Web-based Question Answering Model</title>
    <author>
      <first>Alon</first>
      <last>Talmor</last>
    </author>
    <author>
      <first>Mor</first>
      <last>Geva</last>
    </author>
    <author>
      <first>Jonathan</first>
      <last>Berant</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>161–167</pages>
    <url>http://www.aclweb.org/anthology/S17-1020</url>
    <doi>10.18653/v1/S17-1020</doi>
    <abstract>
      Semantic parsing shines at analyzing complex natural language that
      involves composition and computation over multiple pieces of evidence.
      However, datasets for semantic parsing contain many factoid questions that
      can be answered from a single web document. In this paper, we propose to
      evaluate semantic parsing-based question answering models by comparing
      them to a question answering baseline that queries the web and extracts
      the answer only from web snippets, without access to the target
      knowledge-base. We investigate this approach on COMPLEXQUESTIONS, a
      dataset designed to focus on compositional language, and find that our
      model obtains reasonable performance (∼35 F1 compared to 41 F1 of
      state-of-the-art). We find in our analysis that our model performs well on
      complex questions involving conjunctions, but struggles on questions that
      involve relation composition and superlatives.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>talmor-geva-berant:2017:starSEM</bibkey>
  </paper>
  <paper id='1021'>
    <title>Logical Metonymy in a Distributional Model of Sentence Comprehension</title>
    <author>
      <first>Emmanuele</first>
      <last>Chersoni</last>
    </author>
    <author>
      <first>Alessandro</first>
      <last>Lenci</last>
    </author>
    <author>
      <first>Philippe</first>
      <last>Blache</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>168–177</pages>
    <url>http://www.aclweb.org/anthology/S17-1021</url>
    <doi>10.18653/v1/S17-1021</doi>
    <abstract>
      In theoretical linguistics, logical metonymy is defined as the combination
      of an event-subcategorizing verb with an entity-denoting direct object
      (e.g., The author began the book), so that the interpretation of the VP
      requires the retrieval of a covert event (e.g., writing). Psycholinguistic
      studies have revealed extra processing costs for logical metonymy, a
      phenomenon generally explained with the introduction of new semantic
      structure. In this paper, we present a general distributional model for
      sentence comprehension inspired by the Memory, Unification and Control
      model by Hagoort (2013,2016). We show that our distributional framework
      can account for the extra processing costs of logical metonymy and can
      identify the covert event in a classification task.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>chersoni-lenci-blache:2017:starSEM</bibkey>
  </paper>
  <paper id='1022'>
    <title>Double Trouble: The Problem of Construal in Semantic Annotation of Adpositions</title>
    <author>
      <first>Jena D.</first>
      <last>Hwang</last>
    </author>
    <author>
      <first>Archna</first>
      <last>Bhatia</last>
    </author>
    <author>
      <first>Na-Rae</first>
      <last>Han</last>
    </author>
    <author>
      <first>Tim</first>
      <last>O'Gorman</last>
    </author>
    <author>
      <first>Vivek</first>
      <last>Srikumar</last>
    </author>
    <author>
      <first>Nathan</first>
      <last>Schneider</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>178–188</pages>
    <url>http://www.aclweb.org/anthology/S17-1022</url>
    <attachment type="presentation">S17-1022.Presentation.pdf</attachment>
    <doi>10.18653/v1/S17-1022</doi>
    <abstract>
      We consider the semantics of prepositions, revisiting a broad-coverage
      annotation scheme used for annotating all 4,250 preposition tokens in a
      55,000 word corpus of English. Attempts to apply the scheme to adpositions
      and case markers in other languages, as well as some problematic cases in
      English, have led us to reconsider the assumption that an adposition’s
      lexical contribution is equivalent to the role/relation that it mediates.
      Our proposal is to embrace the potential for construal in adposition use,
      expressing such phenomena directly at the token level to manage complexity
      and avoid sense proliferation. We suggest a framework to represent both
      the scene role and the adposition’s lexical function so they can be
      annotated at scale—supporting automatic, statistical processing of
      domain-general language—and discuss how this representation would allow
      for a simpler inventory of labels.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>hwang-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1023'>
    <title>Issues of Mass and Count: Dealing with `Dual-Life' Nouns</title>
    <author>
      <first>Tibor</first>
      <last>Kiss</last>
    </author>
    <author>
      <first>Francis Jeffry</first>
      <last>Pelletier</last>
    </author>
    <author>
      <first>Halima</first>
      <last>Husic</last>
    </author>
    <author>
      <first>Johanna</first>
      <last>Poppek</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>189–198</pages>
    <url>http://www.aclweb.org/anthology/S17-1023</url>
    <doi>10.18653/v1/S17-1023</doi>
    <abstract>
      The topics of mass and count have been studied for many decades in
      philosophy (e.g., Quine, 1960; Pelletier, 1975), linguistics (e.g.,
      McCawley, 1975; Allen, 1980; Krifka, 1991) and psychology (e.g., Middleton
      et al, 2004; Barner et al, 2009). More recently, interest from within
      computational linguistics has studied the issues involved (e.g.,
      Pustejovsky, 1991; Bond, 2005; Schmidtke &amp; Kuperman, 2016), to name
      just a few. As is pointed out in these works, there are many difficult
      conceptual issues involved in the study of this contrast. In this article
      we study one of these issues – the “Dual-Life” of being simultaneously
      +mass and +count – by means of an unusual combination of human annotation,
      online lexical resources, and online corpora.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>kiss-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1024'>
    <title>Parsing Graphs with Regular Graph Grammars</title>
    <author>
      <first>Sorcha</first>
      <last>Gilroy</last>
    </author>
    <author>
      <first>Adam</first>
      <last>Lopez</last>
    </author>
    <author>
      <first>Sebastian</first>
      <last>Maneth</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>199–208</pages>
    <url>http://www.aclweb.org/anthology/S17-1024</url>
    <doi>10.18653/v1/S17-1024</doi>
    <abstract>
      Recently, several datasets have become available which represent natural
      language phenomena as graphs. Hyperedge Replacement Languages (HRL) have
      been the focus of much attention as a formalism to represent the graphs in
      these datasets. Chiang et al. (2013) prove that HRL graphs can be parsed
      in polynomial time with respect to the size of the input graph. We believe
      that HRL are more expressive than is necessary to represent semantic
      graphs and we propose the use of Regular Graph Languages (RGL; Courcelle
      1991), which is a subfamily of HRL, as a possible alternative. We provide
      a top-down parsing algorithm for RGL that runs in time linear in the size
      of the input graph.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>gilroy-lopez-maneth:2017:starSEM</bibkey>
  </paper>
  <paper id='1025'>
    <title>Embedded Semantic Lexicon Induction with Joint Global and Local Optimization</title>
    <author>
      <first>Sujay Kumar</first>
      <last>Jauhar</last>
    </author>
    <author>
      <first>Eduard</first>
      <last>Hovy</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>209–219</pages>
    <url>http://www.aclweb.org/anthology/S17-1025</url>
    <doi>10.18653/v1/S17-1025</doi>
    <abstract>
      Creating annotated frame lexicons such as PropBank and FrameNet is
      expensive and labor intensive. We present a method to induce an embedded
      frame lexicon in an minimally supervised fashion using nothing more than
      unlabeled predicate-argument word pairs. We hypothesize that aggregating
      such pair selectional preferences across training leads us to a global
      understanding that captures predicate-argument frame structure. Our
      approach revolves around a novel integration between a predictive
      embedding model and an Indian Buffet Process posterior regularizer. We
      show, through our experimental evaluation, that we outperform baselines on
      two tasks and can learn an embedded frame lexicon that is able to capture
      some interesting generalities in relation to hand-crafted semantic frames.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>jauhar-hovy:2017:starSEM</bibkey>
  </paper>
  <paper id='1026'>
    <title>Generating Pattern-Based Entailment Graphs for Relation Extraction</title>
    <author>
      <first>Kathrin</first>
      <last>Eichler</last>
    </author>
    <author>
      <first>Feiyu</first>
      <last>Xu</last>
    </author>
    <author>
      <first>Hans</first>
      <last>Uszkoreit</last>
    </author>
    <author>
      <first>Sebastian</first>
      <last>Krause</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>220–229</pages>
    <url>http://www.aclweb.org/anthology/S17-1026</url>
    <doi>10.18653/v1/S17-1026</doi>
    <abstract>
      Relation extraction is the task of recognizing and extracting relations
      between entities or concepts in texts. A common approach is to exploit
      existing knowledge to learn linguistic patterns expressing the target
      relation and use these patterns for extracting new relation mentions.
      Deriving relation patterns automatically usually results in large numbers
      of candidates, which need to be filtered to derive a subset of patterns
      that reliably extract correct relation mentions. We address the pattern
      selection task by exploiting the knowledge represented by entailment
      graphs, which capture semantic relationships holding among the learned
      pattern candidates. This is motivated by the fact that a pattern may not
      express the target relation explicitly, but still be useful for extracting
      instances for which the relation holds, because its meaning entails the
      meaning of the target relation. We evaluate the usage of both
      automatically generated and gold-standard entailment graphs in a relation
      extraction scenario and present favorable experimental results, exhibiting
      the benefits of structuring and selecting patterns based on entailment
      graphs.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>eichler-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1027'>
    <title>
      Classifying Semantic Clause Types: Modeling Context and Genre
      Characteristics with Recurrent Neural Networks and Attention
    </title>
    <author>
      <first>Maria</first>
      <last>Becker</last>
    </author>
    <author>
      <first>Michael</first>
      <last>Staniek</last>
    </author>
    <author>
      <first>Vivi</first>
      <last>Nastase</last>
    </author>
    <author>
      <first>Alexis</first>
      <last>Palmer</last>
    </author>
    <author>
      <first>Anette</first>
      <last>Frank</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>230–240</pages>
    <url>http://www.aclweb.org/anthology/S17-1027</url>
    <doi>10.18653/v1/S17-1027</doi>
    <abstract>
      Detecting aspectual properties of clauses in the form of situation entity
      types has been shown to depend on a combination of syntactic-semantic and
      contextual features. We explore this task in a deep-learning framework,
      where tuned word representations capture lexical, syntactic and semantic
      features. We introduce an attention mechanism that pinpoints relevant
      context not only for the current instance, but also for the larger
      context. Apart from implicitly capturing task relevant features, the
      advantage of our neural model is that it avoids the need to reproduce
      linguistic features for other languages and is thus more easily
      transferable. We present experiments for English and German that achieve
      competitive performance. We present a novel take on modeling and
      exploiting genre information and showcase the adaptation of our system
      from one language to another.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>becker-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1028'>
    <title>Predictive Linguistic Features of Schizophrenia</title>
    <author>
      <first>Efsun</first>
      <last>Sarioglu Kayi</last>
    </author>
    <author>
      <first>Mona</first>
      <last>Diab</last>
    </author>
    <author>
      <first>Luca</first>
      <last>Pauselli</last>
    </author>
    <author>
      <first>Michael</first>
      <last>Compton</last>
    </author>
    <author>
      <first>Glen</first>
      <last>Coppersmith</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>241–250</pages>
    <url>http://www.aclweb.org/anthology/S17-1028</url>
    <doi>10.18653/v1/S17-1028</doi>
    <abstract>
      Schizophrenia is one of the most disabling and difficult to treat of all
      human medical/health conditions, ranking in the top ten causes of
      disability worldwide. It has been a puzzle in part due to difficulty in
      identifying its basic, fundamental components. Several studies have shown
      that some manifestations of schizophrenia (e.g., the negative symptoms
      that include blunting of speech prosody, as well as the disorganization
      symptoms that lead to disordered language) can be understood from the
      perspective of linguistics. However, schizophrenia research has not kept
      pace with technologies in computational linguistics, especially in
      semantics and pragmatics. As such, we examine the writings of
      schizophrenia patients analyzing their syntax, semantics and pragmatics.
      In addition, we analyze tweets of (self proclaimed) schizophrenia patients
      who publicly discuss their diagnoses. For writing samples dataset,
      syntactic features are found to be the most successful in classification
      whereas for the less structured Twitter dataset, a combination of features
      performed the best.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>sarioglukayi-EtAl:2017:starSEM</bibkey>
  </paper>
  <paper id='1029'>
    <title>
      Learning to Solve Geometry Problems from Natural Language Demonstrations
      in Textbooks
    </title>
    <author>
      <first>Mrinmaya</first>
      <last>Sachan</last>
    </author>
    <author>
      <first>Eric</first>
      <last>Xing</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>251–261</pages>
    <url>http://www.aclweb.org/anthology/S17-1029</url>
    <doi>10.18653/v1/S17-1029</doi>
    <abstract>
      Humans as well as animals are good at imitation. Inspired by this, the
      learning by demonstration view of machine learning learns to perform a
      task from detailed example demonstrations. In this paper, we introduce the
      task of question answering using natural language demonstrations where the
      question answering system is provided with detailed demonstrative
      solutions to questions in natural language. As a case study, we explore
      the task of learning to solve geometry problems using demonstrative
      solutions available in textbooks. We collect a new dataset of
      demonstrative geometry solutions from textbooks and explore approaches
      that learn to interpret these demonstrations as well as to use these
      interpretations to solve geometry problems. Our approaches show
      improvements over the best previously published system for solving
      geometry problems.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>sachan-xing:2017:starSEM</bibkey>
  </paper>
  <paper id='1030'>
    <title>Ways of Asking and Replying in Duplicate Question Detection</title>
    <author>
      <first>João</first>
      <last>António Rodrigues</last>
    </author>
    <author>
      <first>Chakaveh</first>
      <last>Saedi</last>
    </author>
    <author>
      <first>Vladislav</first>
      <last>Maraev</last>
    </author>
    <author>
      <first>João</first>
      <last>Silva</last>
    </author>
    <author>
      <first>António</first>
      <last>Branco</last>
    </author>
    <booktitle>
      Proceedings of the 6th Joint Conference on Lexical and Computational
      Semantics (*SEM 2017)
    </booktitle>
    <month>August</month>
    <year>2017</year>
    <address>Vancouver, Canada</address>
    <publisher>Association for Computational Linguistics</publisher>
    <pages>262–270</pages>
    <url>http://www.aclweb.org/anthology/S17-1030</url>
    <doi>10.18653/v1/S17-1030</doi>
    <abstract>
      This paper presents the results of systematic experimentation on the
      impact in duplicate question detection of different types of questions
      across both a number of established approaches and a novel, superior one
      used to address this language processing task. This study permits to gain
      a novel insight on the different levels of robustness of the diverse
      detection methods with respect to different conditions of their
      application, including the ones that approximate real usage scenarios.
    </abstract>
    <bibtype>inproceedings</bibtype>
    <bibkey>antoniorodrigues-EtAl:2017:starSEM</bibkey>
  </paper>
</volume>