@inproceedings{cao-etal-2020-pretrained,
title = "Pretrained Language Models for Dialogue Generation with Multiple Input Sources",
author = "Cao, Yu and
Bi, Wei and
Fang, Meng and
Tao, Dacheng",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2020",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2020.findings-emnlp.81",
doi = "10.18653/v1/2020.findings-emnlp.81",
pages = "909--917",
abstract = "Large-scale pretrained language models have achieved outstanding performance on natural language understanding tasks. However, it is still under investigating how to apply them to dialogue generation tasks, especially those with responses conditioned on multiple sources. Previous work simply concatenates all input sources or averages information from different input sources. In this work, we study dialogue models with multiple input sources adapted from the pretrained language model GPT2. We explore various methods to fuse multiple separate attention information corresponding to different sources. Our experimental results show that proper fusion methods deliver higher relevance with dialogue history than simple fusion baselines.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="cao-etal-2020-pretrained">
<titleInfo>
<title>Pretrained Language Models for Dialogue Generation with Multiple Input Sources</title>
</titleInfo>
<name type="personal">
<namePart type="given">Yu</namePart>
<namePart type="family">Cao</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Wei</namePart>
<namePart type="family">Bi</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Meng</namePart>
<namePart type="family">Fang</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Dacheng</namePart>
<namePart type="family">Tao</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2020-11</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Findings of the Association for Computational Linguistics: EMNLP 2020</title>
</titleInfo>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Online</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Large-scale pretrained language models have achieved outstanding performance on natural language understanding tasks. However, it is still under investigating how to apply them to dialogue generation tasks, especially those with responses conditioned on multiple sources. Previous work simply concatenates all input sources or averages information from different input sources. In this work, we study dialogue models with multiple input sources adapted from the pretrained language model GPT2. We explore various methods to fuse multiple separate attention information corresponding to different sources. Our experimental results show that proper fusion methods deliver higher relevance with dialogue history than simple fusion baselines.</abstract>
<identifier type="citekey">cao-etal-2020-pretrained</identifier>
<identifier type="doi">10.18653/v1/2020.findings-emnlp.81</identifier>
<location>
<url>https://aclanthology.org/2020.findings-emnlp.81</url>
</location>
<part>
<date>2020-11</date>
<extent unit="page">
<start>909</start>
<end>917</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Pretrained Language Models for Dialogue Generation with Multiple Input Sources
%A Cao, Yu
%A Bi, Wei
%A Fang, Meng
%A Tao, Dacheng
%S Findings of the Association for Computational Linguistics: EMNLP 2020
%D 2020
%8 November
%I Association for Computational Linguistics
%C Online
%F cao-etal-2020-pretrained
%X Large-scale pretrained language models have achieved outstanding performance on natural language understanding tasks. However, it is still under investigating how to apply them to dialogue generation tasks, especially those with responses conditioned on multiple sources. Previous work simply concatenates all input sources or averages information from different input sources. In this work, we study dialogue models with multiple input sources adapted from the pretrained language model GPT2. We explore various methods to fuse multiple separate attention information corresponding to different sources. Our experimental results show that proper fusion methods deliver higher relevance with dialogue history than simple fusion baselines.
%R 10.18653/v1/2020.findings-emnlp.81
%U https://aclanthology.org/2020.findings-emnlp.81
%U https://doi.org/10.18653/v1/2020.findings-emnlp.81
%P 909-917
Markdown (Informal)
[Pretrained Language Models for Dialogue Generation with Multiple Input Sources](https://aclanthology.org/2020.findings-emnlp.81) (Cao et al., Findings 2020)
ACL