@inproceedings{oliveira-2023-acquisition,
title = "On the Acquisition of {W}ord{N}et Relations in {P}ortuguese from Pretrained Masked Language Models",
author = "Oliveira, Hugo Gon{\c{c}}alo",
editor = "Rigau, German and
Bond, Francis and
Rademaker, Alexandre",
booktitle = "Proceedings of the 12th Global Wordnet Conference",
month = jan,
year = "2023",
address = "University of the Basque Country, Donostia - San Sebastian, Basque Country",
publisher = "Global Wordnet Association",
url = "https://aclanthology.org/2023.gwc-1.5",
pages = "41--49",
abstract = "This paper studies the application of pretrained BERT in the acquisition of synonyms, antonyms, hypernyms and hyponyms in Portuguese. Masked patterns indicating those relations were compiled with the help of a service for validating semantic relations, and then used for prompting three pretrained BERT models, one multilingual and two for Portuguese (base and large). Predictions for the masks were evaluated in two different test sets. Results achieved by the monolingual models are interesting enough for considering these models as a source for enriching wordnets, especially when predicting hypernyms of nouns. Previously reported performances on prediction were improved with new patterns and with the large model. When it comes to selecting the related word from a set of four options, performance is even better, but not enough for outperforming the selection of the most similar word, as computed with static word embeddings.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="oliveira-2023-acquisition">
<titleInfo>
<title>On the Acquisition of WordNet Relations in Portuguese from Pretrained Masked Language Models</title>
</titleInfo>
<name type="personal">
<namePart type="given">Hugo</namePart>
<namePart type="given">Gonçalo</namePart>
<namePart type="family">Oliveira</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2023-01</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 12th Global Wordnet Conference</title>
</titleInfo>
<name type="personal">
<namePart type="given">German</namePart>
<namePart type="family">Rigau</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Francis</namePart>
<namePart type="family">Bond</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Alexandre</namePart>
<namePart type="family">Rademaker</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Global Wordnet Association</publisher>
<place>
<placeTerm type="text">University of the Basque Country, Donostia - San Sebastian, Basque Country</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>This paper studies the application of pretrained BERT in the acquisition of synonyms, antonyms, hypernyms and hyponyms in Portuguese. Masked patterns indicating those relations were compiled with the help of a service for validating semantic relations, and then used for prompting three pretrained BERT models, one multilingual and two for Portuguese (base and large). Predictions for the masks were evaluated in two different test sets. Results achieved by the monolingual models are interesting enough for considering these models as a source for enriching wordnets, especially when predicting hypernyms of nouns. Previously reported performances on prediction were improved with new patterns and with the large model. When it comes to selecting the related word from a set of four options, performance is even better, but not enough for outperforming the selection of the most similar word, as computed with static word embeddings.</abstract>
<identifier type="citekey">oliveira-2023-acquisition</identifier>
<location>
<url>https://aclanthology.org/2023.gwc-1.5</url>
</location>
<part>
<date>2023-01</date>
<extent unit="page">
<start>41</start>
<end>49</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T On the Acquisition of WordNet Relations in Portuguese from Pretrained Masked Language Models
%A Oliveira, Hugo Gonçalo
%Y Rigau, German
%Y Bond, Francis
%Y Rademaker, Alexandre
%S Proceedings of the 12th Global Wordnet Conference
%D 2023
%8 January
%I Global Wordnet Association
%C University of the Basque Country, Donostia - San Sebastian, Basque Country
%F oliveira-2023-acquisition
%X This paper studies the application of pretrained BERT in the acquisition of synonyms, antonyms, hypernyms and hyponyms in Portuguese. Masked patterns indicating those relations were compiled with the help of a service for validating semantic relations, and then used for prompting three pretrained BERT models, one multilingual and two for Portuguese (base and large). Predictions for the masks were evaluated in two different test sets. Results achieved by the monolingual models are interesting enough for considering these models as a source for enriching wordnets, especially when predicting hypernyms of nouns. Previously reported performances on prediction were improved with new patterns and with the large model. When it comes to selecting the related word from a set of four options, performance is even better, but not enough for outperforming the selection of the most similar word, as computed with static word embeddings.
%U https://aclanthology.org/2023.gwc-1.5
%P 41-49
Markdown (Informal)
[On the Acquisition of WordNet Relations in Portuguese from Pretrained Masked Language Models](https://aclanthology.org/2023.gwc-1.5) (Oliveira, GWC 2023)
ACL