@article{cheng-etal-2019-learning,
title = "Learning an Executable Neural Semantic Parser",
author = "Cheng, Jianpeng and
Reddy, Siva and
Saraswat, Vijay and
Lapata, Mirella",
journal = "Computational Linguistics",
volume = "45",
number = "1",
month = mar,
year = "2019",
address = "Cambridge, MA",
publisher = "MIT Press",
url = "https://aclanthology.org/J19-1002",
doi = "10.1162/coli_a_00342",
pages = "59--94",
abstract = "This article describes a neural semantic parser that maps natural language utterances onto logical forms that can be executed against a task-specific environment, such as a knowledge base or a database, to produce a response. The parser generates tree-structured logical forms with a transition-based approach, combining a generic tree-generation algorithm with domain-general grammar defined by the logical language. The generation process is modeled by structured recurrent neural networks, which provide a rich encoding of the sentential context and generation history for making predictions. To tackle mismatches between natural language and logical form tokens, various attention mechanisms are explored. Finally, we consider different training settings for the neural semantic parser, including fully supervised training where annotated logical forms are given, weakly supervised training where denotations are provided, and distant supervision where only unlabeled sentences and a knowledge base are available. Experiments across a wide range of data sets demonstrate the effectiveness of our parser.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="cheng-etal-2019-learning">
<titleInfo>
<title>Learning an Executable Neural Semantic Parser</title>
</titleInfo>
<name type="personal">
<namePart type="given">Jianpeng</namePart>
<namePart type="family">Cheng</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Siva</namePart>
<namePart type="family">Reddy</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Vijay</namePart>
<namePart type="family">Saraswat</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Mirella</namePart>
<namePart type="family">Lapata</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2019-03</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<genre authority="bibutilsgt">journal article</genre>
<relatedItem type="host">
<titleInfo>
<title>Computational Linguistics</title>
</titleInfo>
<originInfo>
<issuance>continuing</issuance>
<publisher>MIT Press</publisher>
<place>
<placeTerm type="text">Cambridge, MA</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">periodical</genre>
<genre authority="bibutilsgt">academic journal</genre>
</relatedItem>
<abstract>This article describes a neural semantic parser that maps natural language utterances onto logical forms that can be executed against a task-specific environment, such as a knowledge base or a database, to produce a response. The parser generates tree-structured logical forms with a transition-based approach, combining a generic tree-generation algorithm with domain-general grammar defined by the logical language. The generation process is modeled by structured recurrent neural networks, which provide a rich encoding of the sentential context and generation history for making predictions. To tackle mismatches between natural language and logical form tokens, various attention mechanisms are explored. Finally, we consider different training settings for the neural semantic parser, including fully supervised training where annotated logical forms are given, weakly supervised training where denotations are provided, and distant supervision where only unlabeled sentences and a knowledge base are available. Experiments across a wide range of data sets demonstrate the effectiveness of our parser.</abstract>
<identifier type="citekey">cheng-etal-2019-learning</identifier>
<identifier type="doi">10.1162/coli_a_00342</identifier>
<location>
<url>https://aclanthology.org/J19-1002</url>
</location>
<part>
<date>2019-03</date>
<detail type="volume"><number>45</number></detail>
<detail type="issue"><number>1</number></detail>
<extent unit="page">
<start>59</start>
<end>94</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Journal Article
%T Learning an Executable Neural Semantic Parser
%A Cheng, Jianpeng
%A Reddy, Siva
%A Saraswat, Vijay
%A Lapata, Mirella
%J Computational Linguistics
%D 2019
%8 March
%V 45
%N 1
%I MIT Press
%C Cambridge, MA
%F cheng-etal-2019-learning
%X This article describes a neural semantic parser that maps natural language utterances onto logical forms that can be executed against a task-specific environment, such as a knowledge base or a database, to produce a response. The parser generates tree-structured logical forms with a transition-based approach, combining a generic tree-generation algorithm with domain-general grammar defined by the logical language. The generation process is modeled by structured recurrent neural networks, which provide a rich encoding of the sentential context and generation history for making predictions. To tackle mismatches between natural language and logical form tokens, various attention mechanisms are explored. Finally, we consider different training settings for the neural semantic parser, including fully supervised training where annotated logical forms are given, weakly supervised training where denotations are provided, and distant supervision where only unlabeled sentences and a knowledge base are available. Experiments across a wide range of data sets demonstrate the effectiveness of our parser.
%R 10.1162/coli_a_00342
%U https://aclanthology.org/J19-1002
%U https://doi.org/10.1162/coli_a_00342
%P 59-94
Markdown (Informal)
[Learning an Executable Neural Semantic Parser](https://aclanthology.org/J19-1002) (Cheng et al., CL 2019)
ACL