@inproceedings{larsson-etal-2023-ttr,
title = "{TTR} at the {SPA}: Relating type-theoretical semantics to neural semantic pointers",
author = "Larsson, Staffan and
Cooper, Robin and
Ginzburg, Jonathan and
Luecking, Andy",
editor = "Chatzikyriakidis, Stergios and
de Paiva, Valeria",
booktitle = "Proceedings of the 4th Natural Logic Meets Machine Learning Workshop",
month = jun,
year = "2023",
address = "Nancy, France",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.naloma-1.5",
pages = "41--50",
abstract = "This paper considers how the kind of formal semantic objects used in TTR (a theory of types with records, Cooper 2013) might be related to the vector representations used in Eliasmith (2013). An advantage of doing this is that it would immediately give us a neural representation for TTR objects as Eliasmith relates vectors to neural activity in his semantic pointer architecture (SPA). This would be an alternative using convolution to the suggestions made by Cooper (2019) based on the phasing of neural activity. The project seems potentially hopeful since all complex TTR objects are constructed from labelled sets (essentially sets of ordered pairs consisting of labels and values) which might be seen as corresponding to the representation of structured objects which Eliasmith achieves using superposition and circular convolution.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="larsson-etal-2023-ttr">
<titleInfo>
<title>TTR at the SPA: Relating type-theoretical semantics to neural semantic pointers</title>
</titleInfo>
<name type="personal">
<namePart type="given">Staffan</namePart>
<namePart type="family">Larsson</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Robin</namePart>
<namePart type="family">Cooper</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jonathan</namePart>
<namePart type="family">Ginzburg</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Andy</namePart>
<namePart type="family">Luecking</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2023-06</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 4th Natural Logic Meets Machine Learning Workshop</title>
</titleInfo>
<name type="personal">
<namePart type="given">Stergios</namePart>
<namePart type="family">Chatzikyriakidis</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Valeria</namePart>
<namePart type="family">de Paiva</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Nancy, France</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>This paper considers how the kind of formal semantic objects used in TTR (a theory of types with records, Cooper 2013) might be related to the vector representations used in Eliasmith (2013). An advantage of doing this is that it would immediately give us a neural representation for TTR objects as Eliasmith relates vectors to neural activity in his semantic pointer architecture (SPA). This would be an alternative using convolution to the suggestions made by Cooper (2019) based on the phasing of neural activity. The project seems potentially hopeful since all complex TTR objects are constructed from labelled sets (essentially sets of ordered pairs consisting of labels and values) which might be seen as corresponding to the representation of structured objects which Eliasmith achieves using superposition and circular convolution.</abstract>
<identifier type="citekey">larsson-etal-2023-ttr</identifier>
<location>
<url>https://aclanthology.org/2023.naloma-1.5</url>
</location>
<part>
<date>2023-06</date>
<extent unit="page">
<start>41</start>
<end>50</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T TTR at the SPA: Relating type-theoretical semantics to neural semantic pointers
%A Larsson, Staffan
%A Cooper, Robin
%A Ginzburg, Jonathan
%A Luecking, Andy
%Y Chatzikyriakidis, Stergios
%Y de Paiva, Valeria
%S Proceedings of the 4th Natural Logic Meets Machine Learning Workshop
%D 2023
%8 June
%I Association for Computational Linguistics
%C Nancy, France
%F larsson-etal-2023-ttr
%X This paper considers how the kind of formal semantic objects used in TTR (a theory of types with records, Cooper 2013) might be related to the vector representations used in Eliasmith (2013). An advantage of doing this is that it would immediately give us a neural representation for TTR objects as Eliasmith relates vectors to neural activity in his semantic pointer architecture (SPA). This would be an alternative using convolution to the suggestions made by Cooper (2019) based on the phasing of neural activity. The project seems potentially hopeful since all complex TTR objects are constructed from labelled sets (essentially sets of ordered pairs consisting of labels and values) which might be seen as corresponding to the representation of structured objects which Eliasmith achieves using superposition and circular convolution.
%U https://aclanthology.org/2023.naloma-1.5
%P 41-50
Markdown (Informal)
[TTR at the SPA: Relating type-theoretical semantics to neural semantic pointers](https://aclanthology.org/2023.naloma-1.5) (Larsson et al., NALOMA-WS 2023)
ACL