@inproceedings{arora-etal-2022-exposure,
title = "Why Exposure Bias Matters: An Imitation Learning Perspective of Error Accumulation in Language Generation",
author = "Arora, Kushal and
El Asri, Layla and
Bahuleyan, Hareesh and
Cheung, Jackie",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2022",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.findings-acl.58",
doi = "10.18653/v1/2022.findings-acl.58",
pages = "700--710",
abstract = "Current language generation models suffer from issues such as repetition, incoherence, and hallucinations. An often-repeated hypothesis for this brittleness of generation models is that it is caused by the training and the generation procedure mismatch, also referred to as exposure bias. In this paper, we verify this hypothesis by analyzing exposure bias from an imitation learning perspective. We show that exposure bias leads to an accumulation of errors during generation, analyze why perplexity fails to capture this accumulation of errors, and empirically show that this accumulation results in poor generation quality.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="arora-etal-2022-exposure">
<titleInfo>
<title>Why Exposure Bias Matters: An Imitation Learning Perspective of Error Accumulation in Language Generation</title>
</titleInfo>
<name type="personal">
<namePart type="given">Kushal</namePart>
<namePart type="family">Arora</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Layla</namePart>
<namePart type="family">El Asri</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Hareesh</namePart>
<namePart type="family">Bahuleyan</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jackie</namePart>
<namePart type="family">Cheung</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2022-05</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Findings of the Association for Computational Linguistics: ACL 2022</title>
</titleInfo>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Dublin, Ireland</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Current language generation models suffer from issues such as repetition, incoherence, and hallucinations. An often-repeated hypothesis for this brittleness of generation models is that it is caused by the training and the generation procedure mismatch, also referred to as exposure bias. In this paper, we verify this hypothesis by analyzing exposure bias from an imitation learning perspective. We show that exposure bias leads to an accumulation of errors during generation, analyze why perplexity fails to capture this accumulation of errors, and empirically show that this accumulation results in poor generation quality.</abstract>
<identifier type="citekey">arora-etal-2022-exposure</identifier>
<identifier type="doi">10.18653/v1/2022.findings-acl.58</identifier>
<location>
<url>https://aclanthology.org/2022.findings-acl.58</url>
</location>
<part>
<date>2022-05</date>
<extent unit="page">
<start>700</start>
<end>710</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Why Exposure Bias Matters: An Imitation Learning Perspective of Error Accumulation in Language Generation
%A Arora, Kushal
%A El Asri, Layla
%A Bahuleyan, Hareesh
%A Cheung, Jackie
%S Findings of the Association for Computational Linguistics: ACL 2022
%D 2022
%8 May
%I Association for Computational Linguistics
%C Dublin, Ireland
%F arora-etal-2022-exposure
%X Current language generation models suffer from issues such as repetition, incoherence, and hallucinations. An often-repeated hypothesis for this brittleness of generation models is that it is caused by the training and the generation procedure mismatch, also referred to as exposure bias. In this paper, we verify this hypothesis by analyzing exposure bias from an imitation learning perspective. We show that exposure bias leads to an accumulation of errors during generation, analyze why perplexity fails to capture this accumulation of errors, and empirically show that this accumulation results in poor generation quality.
%R 10.18653/v1/2022.findings-acl.58
%U https://aclanthology.org/2022.findings-acl.58
%U https://doi.org/10.18653/v1/2022.findings-acl.58
%P 700-710
Markdown (Informal)
[Why Exposure Bias Matters: An Imitation Learning Perspective of Error Accumulation in Language Generation](https://aclanthology.org/2022.findings-acl.58) (Arora et al., Findings 2022)
ACL