@inproceedings{kimura-etal-2021-towards-language,
title = "Towards a Language Model for Temporal Commonsense Reasoning",
author = "Kimura, Mayuko and
Kanashiro Pereira, Lis and
Kobayashi, Ichiro",
editor = "Djabri, Souhila and
Gimadi, Dinara and
Mihaylova, Tsvetomila and
Nikolova-Koleva, Ivelina",
booktitle = "Proceedings of the Student Research Workshop Associated with RANLP 2021",
month = sep,
year = "2021",
address = "Online",
publisher = "INCOMA Ltd.",
url = "https://aclanthology.org/2021.ranlp-srw.12",
pages = "78--84",
abstract = "Temporal commonsense reasoning is a challenging task as it requires temporal knowledge usually not explicit in text. In this work, we propose an ensemble model for temporal commonsense reasoning. Our model relies on pre-trained contextual representations from transformer-based language models (i.e., BERT), and on a variety of training methods for enhancing model generalization: 1) multi-step fine-tuning using carefully selected auxiliary tasks and datasets, and 2) a specifically designed temporal masked language model task aimed to capture temporal commonsense knowledge. Our model greatly outperforms the standard fine-tuning approach and strong baselines on the MC-TACO dataset.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="kimura-etal-2021-towards-language">
<titleInfo>
<title>Towards a Language Model for Temporal Commonsense Reasoning</title>
</titleInfo>
<name type="personal">
<namePart type="given">Mayuko</namePart>
<namePart type="family">Kimura</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Lis</namePart>
<namePart type="family">Kanashiro Pereira</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Ichiro</namePart>
<namePart type="family">Kobayashi</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2021-09</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Student Research Workshop Associated with RANLP 2021</title>
</titleInfo>
<name type="personal">
<namePart type="given">Souhila</namePart>
<namePart type="family">Djabri</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Dinara</namePart>
<namePart type="family">Gimadi</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Tsvetomila</namePart>
<namePart type="family">Mihaylova</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Ivelina</namePart>
<namePart type="family">Nikolova-Koleva</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>INCOMA Ltd.</publisher>
<place>
<placeTerm type="text">Online</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Temporal commonsense reasoning is a challenging task as it requires temporal knowledge usually not explicit in text. In this work, we propose an ensemble model for temporal commonsense reasoning. Our model relies on pre-trained contextual representations from transformer-based language models (i.e., BERT), and on a variety of training methods for enhancing model generalization: 1) multi-step fine-tuning using carefully selected auxiliary tasks and datasets, and 2) a specifically designed temporal masked language model task aimed to capture temporal commonsense knowledge. Our model greatly outperforms the standard fine-tuning approach and strong baselines on the MC-TACO dataset.</abstract>
<identifier type="citekey">kimura-etal-2021-towards-language</identifier>
<location>
<url>https://aclanthology.org/2021.ranlp-srw.12</url>
</location>
<part>
<date>2021-09</date>
<extent unit="page">
<start>78</start>
<end>84</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Towards a Language Model for Temporal Commonsense Reasoning
%A Kimura, Mayuko
%A Kanashiro Pereira, Lis
%A Kobayashi, Ichiro
%Y Djabri, Souhila
%Y Gimadi, Dinara
%Y Mihaylova, Tsvetomila
%Y Nikolova-Koleva, Ivelina
%S Proceedings of the Student Research Workshop Associated with RANLP 2021
%D 2021
%8 September
%I INCOMA Ltd.
%C Online
%F kimura-etal-2021-towards-language
%X Temporal commonsense reasoning is a challenging task as it requires temporal knowledge usually not explicit in text. In this work, we propose an ensemble model for temporal commonsense reasoning. Our model relies on pre-trained contextual representations from transformer-based language models (i.e., BERT), and on a variety of training methods for enhancing model generalization: 1) multi-step fine-tuning using carefully selected auxiliary tasks and datasets, and 2) a specifically designed temporal masked language model task aimed to capture temporal commonsense knowledge. Our model greatly outperforms the standard fine-tuning approach and strong baselines on the MC-TACO dataset.
%U https://aclanthology.org/2021.ranlp-srw.12
%P 78-84
Markdown (Informal)
[Towards a Language Model for Temporal Commonsense Reasoning](https://aclanthology.org/2021.ranlp-srw.12) (Kimura et al., RANLP 2021)
ACL