@inproceedings{barker-etal-2021-ibm,
title = "{IBM} {MNLP} {IE} at {CASE} 2021 Task 2: {NLI} Reranking for Zero-Shot Text Classification",
author = "Barker, Ken and
Awasthy, Parul and
Ni, Jian and
Florian, Radu",
editor = {H{\"u}rriyeto{\u{g}}lu, Ali},
booktitle = "Proceedings of the 4th Workshop on Challenges and Applications of Automated Extraction of Socio-political Events from Text (CASE 2021)",
month = aug,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.case-1.24/",
doi = "10.18653/v1/2021.case-1.24",
pages = "193--202",
abstract = "Supervised models can achieve very high accuracy for fine-grained text classification. In practice, however, training data may be abundant for some types but scarce or even non-existent for others. We propose a hybrid architecture that uses as much labeled data as available for fine-tuning classification models, while also allowing for types with little (few-shot) or no (zero-shot) labeled data. In particular, we pair a supervised text classification model with a Natural Language Inference (NLI) reranking model. The NLI reranker uses a textual representation of target types that allows it to score the strength with which a type is implied by a text, without requiring training data for the types. Experiments show that the NLI model is very sensitive to the choice of textual representation, but can be effective for classifying unseen types. It can also improve classification accuracy for the known types of an already highly accurate supervised model."
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="barker-etal-2021-ibm">
<titleInfo>
<title>IBM MNLP IE at CASE 2021 Task 2: NLI Reranking for Zero-Shot Text Classification</title>
</titleInfo>
<name type="personal">
<namePart type="given">Ken</namePart>
<namePart type="family">Barker</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Parul</namePart>
<namePart type="family">Awasthy</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jian</namePart>
<namePart type="family">Ni</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Radu</namePart>
<namePart type="family">Florian</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2021-08</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 4th Workshop on Challenges and Applications of Automated Extraction of Socio-political Events from Text (CASE 2021)</title>
</titleInfo>
<name type="personal">
<namePart type="given">Ali</namePart>
<namePart type="family">Hürriyetoğlu</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Online</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Supervised models can achieve very high accuracy for fine-grained text classification. In practice, however, training data may be abundant for some types but scarce or even non-existent for others. We propose a hybrid architecture that uses as much labeled data as available for fine-tuning classification models, while also allowing for types with little (few-shot) or no (zero-shot) labeled data. In particular, we pair a supervised text classification model with a Natural Language Inference (NLI) reranking model. The NLI reranker uses a textual representation of target types that allows it to score the strength with which a type is implied by a text, without requiring training data for the types. Experiments show that the NLI model is very sensitive to the choice of textual representation, but can be effective for classifying unseen types. It can also improve classification accuracy for the known types of an already highly accurate supervised model.</abstract>
<identifier type="citekey">barker-etal-2021-ibm</identifier>
<identifier type="doi">10.18653/v1/2021.case-1.24</identifier>
<location>
<url>https://aclanthology.org/2021.case-1.24/</url>
</location>
<part>
<date>2021-08</date>
<extent unit="page">
<start>193</start>
<end>202</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T IBM MNLP IE at CASE 2021 Task 2: NLI Reranking for Zero-Shot Text Classification
%A Barker, Ken
%A Awasthy, Parul
%A Ni, Jian
%A Florian, Radu
%Y Hürriyetoğlu, Ali
%S Proceedings of the 4th Workshop on Challenges and Applications of Automated Extraction of Socio-political Events from Text (CASE 2021)
%D 2021
%8 August
%I Association for Computational Linguistics
%C Online
%F barker-etal-2021-ibm
%X Supervised models can achieve very high accuracy for fine-grained text classification. In practice, however, training data may be abundant for some types but scarce or even non-existent for others. We propose a hybrid architecture that uses as much labeled data as available for fine-tuning classification models, while also allowing for types with little (few-shot) or no (zero-shot) labeled data. In particular, we pair a supervised text classification model with a Natural Language Inference (NLI) reranking model. The NLI reranker uses a textual representation of target types that allows it to score the strength with which a type is implied by a text, without requiring training data for the types. Experiments show that the NLI model is very sensitive to the choice of textual representation, but can be effective for classifying unseen types. It can also improve classification accuracy for the known types of an already highly accurate supervised model.
%R 10.18653/v1/2021.case-1.24
%U https://aclanthology.org/2021.case-1.24/
%U https://doi.org/10.18653/v1/2021.case-1.24
%P 193-202
Markdown (Informal)
[IBM MNLP IE at CASE 2021 Task 2: NLI Reranking for Zero-Shot Text Classification](https://aclanthology.org/2021.case-1.24/) (Barker et al., CASE 2021)
ACL