@inproceedings{chen-zhang-2020-representing,
title = "Representing and Reconstructing {P}hy{SH}: Which Embedding Competent?",
author = "Chen, Xiaoli and
Zhang, Zhixiong",
editor = "Knoth, Petr and
Stahl, Christopher and
Gyawali, Bikash and
Pride, David and
Kunnath, Suchetha N. and
Herrmannova, Drahomira",
booktitle = "Proceedings of the 8th International Workshop on Mining Scientific Publications",
month = "05 " # aug,
year = "2020",
address = "Wuhan, China",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2020.wosp-1.7",
pages = "48--53",
abstract = "Recent advances in natural language processing make embedding representations dominate the computing language world. Though it is taken for granted, we actually have limited knowledge of how these embeddings perform in representing the complex hierarchy of domain scientific knowledge. In this paper, we conduct a comprehensive comparison of well-known embeddings{'} capability in capturing the hierarchical Physics knowledge. Several key findings are: i, Poincar{\'e} embeddings do outperform if trained on PhySH taxonomy, but it fails if trained on co-occurrence pairs which are extracted from raw text. ii, No algorithm can properly learn hierarchies from the more realistic case of co-occurrence pairs, which contains more noisy relations other than hierarchical relations. iii, Our statistic analysis of Poincar{\'e} embedding{'}s representation of PhySH shows successful hierarchical representation share two characteristics: firstly, upper-level terms have a smaller semantic distance to root; secondly, upper-level hypernym-hyponym pairs should be further apart than lower-level hypernym-hyponym pairs.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="chen-zhang-2020-representing">
<titleInfo>
<title>Representing and Reconstructing PhySH: Which Embedding Competent?</title>
</titleInfo>
<name type="personal">
<namePart type="given">Xiaoli</namePart>
<namePart type="family">Chen</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Zhixiong</namePart>
<namePart type="family">Zhang</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2020-05 aug</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 8th International Workshop on Mining Scientific Publications</title>
</titleInfo>
<name type="personal">
<namePart type="given">Petr</namePart>
<namePart type="family">Knoth</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Christopher</namePart>
<namePart type="family">Stahl</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Bikash</namePart>
<namePart type="family">Gyawali</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">David</namePart>
<namePart type="family">Pride</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Suchetha</namePart>
<namePart type="given">N</namePart>
<namePart type="family">Kunnath</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Drahomira</namePart>
<namePart type="family">Herrmannova</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Wuhan, China</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Recent advances in natural language processing make embedding representations dominate the computing language world. Though it is taken for granted, we actually have limited knowledge of how these embeddings perform in representing the complex hierarchy of domain scientific knowledge. In this paper, we conduct a comprehensive comparison of well-known embeddings’ capability in capturing the hierarchical Physics knowledge. Several key findings are: i, Poincaré embeddings do outperform if trained on PhySH taxonomy, but it fails if trained on co-occurrence pairs which are extracted from raw text. ii, No algorithm can properly learn hierarchies from the more realistic case of co-occurrence pairs, which contains more noisy relations other than hierarchical relations. iii, Our statistic analysis of Poincaré embedding’s representation of PhySH shows successful hierarchical representation share two characteristics: firstly, upper-level terms have a smaller semantic distance to root; secondly, upper-level hypernym-hyponym pairs should be further apart than lower-level hypernym-hyponym pairs.</abstract>
<identifier type="citekey">chen-zhang-2020-representing</identifier>
<location>
<url>https://aclanthology.org/2020.wosp-1.7</url>
</location>
<part>
<date>2020-05 aug</date>
<extent unit="page">
<start>48</start>
<end>53</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Representing and Reconstructing PhySH: Which Embedding Competent?
%A Chen, Xiaoli
%A Zhang, Zhixiong
%Y Knoth, Petr
%Y Stahl, Christopher
%Y Gyawali, Bikash
%Y Pride, David
%Y Kunnath, Suchetha N.
%Y Herrmannova, Drahomira
%S Proceedings of the 8th International Workshop on Mining Scientific Publications
%D 2020
%8 05 aug
%I Association for Computational Linguistics
%C Wuhan, China
%F chen-zhang-2020-representing
%X Recent advances in natural language processing make embedding representations dominate the computing language world. Though it is taken for granted, we actually have limited knowledge of how these embeddings perform in representing the complex hierarchy of domain scientific knowledge. In this paper, we conduct a comprehensive comparison of well-known embeddings’ capability in capturing the hierarchical Physics knowledge. Several key findings are: i, Poincaré embeddings do outperform if trained on PhySH taxonomy, but it fails if trained on co-occurrence pairs which are extracted from raw text. ii, No algorithm can properly learn hierarchies from the more realistic case of co-occurrence pairs, which contains more noisy relations other than hierarchical relations. iii, Our statistic analysis of Poincaré embedding’s representation of PhySH shows successful hierarchical representation share two characteristics: firstly, upper-level terms have a smaller semantic distance to root; secondly, upper-level hypernym-hyponym pairs should be further apart than lower-level hypernym-hyponym pairs.
%U https://aclanthology.org/2020.wosp-1.7
%P 48-53
Markdown (Informal)
[Representing and Reconstructing PhySH: Which Embedding Competent?](https://aclanthology.org/2020.wosp-1.7) (Chen & Zhang, WOSP 2020)
ACL