@inproceedings{wang-etal-2023-pre-trained,
title = "Pre-trained Model In {A}ncient-{C}hinese-to-{M}odern-{C}hinese Machine Translation",
author = "Wang, Jiahui and
Zhang, Xuqin and
Li, Jiahuan and
Huang, Shujian",
booktitle = "Proceedings of ALT2023: Ancient Language Translation Workshop",
month = sep,
year = "2023",
address = "Macau SAR, China",
publisher = "Asia-Pacific Association for Machine Translation",
url = "https://aclanthology.org/2023.alt-1.3",
pages = "23--28",
abstract = "This paper presents an analysis of the pre-trained Transformer model Neural Machine Translation (NMT) for the Ancient-Chinese-to-Modern-Chinese machine translation task.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="wang-etal-2023-pre-trained">
<titleInfo>
<title>Pre-trained Model In Ancient-Chinese-to-Modern-Chinese Machine Translation</title>
</titleInfo>
<name type="personal">
<namePart type="given">Jiahui</namePart>
<namePart type="family">Wang</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Xuqin</namePart>
<namePart type="family">Zhang</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jiahuan</namePart>
<namePart type="family">Li</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Shujian</namePart>
<namePart type="family">Huang</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2023-09</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of ALT2023: Ancient Language Translation Workshop</title>
</titleInfo>
<originInfo>
<publisher>Asia-Pacific Association for Machine Translation</publisher>
<place>
<placeTerm type="text">Macau SAR, China</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>This paper presents an analysis of the pre-trained Transformer model Neural Machine Translation (NMT) for the Ancient-Chinese-to-Modern-Chinese machine translation task.</abstract>
<identifier type="citekey">wang-etal-2023-pre-trained</identifier>
<location>
<url>https://aclanthology.org/2023.alt-1.3</url>
</location>
<part>
<date>2023-09</date>
<extent unit="page">
<start>23</start>
<end>28</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Pre-trained Model In Ancient-Chinese-to-Modern-Chinese Machine Translation
%A Wang, Jiahui
%A Zhang, Xuqin
%A Li, Jiahuan
%A Huang, Shujian
%S Proceedings of ALT2023: Ancient Language Translation Workshop
%D 2023
%8 September
%I Asia-Pacific Association for Machine Translation
%C Macau SAR, China
%F wang-etal-2023-pre-trained
%X This paper presents an analysis of the pre-trained Transformer model Neural Machine Translation (NMT) for the Ancient-Chinese-to-Modern-Chinese machine translation task.
%U https://aclanthology.org/2023.alt-1.3
%P 23-28
Markdown (Informal)
[Pre-trained Model In Ancient-Chinese-to-Modern-Chinese Machine Translation](https://aclanthology.org/2023.alt-1.3) (Wang et al., alt 2023)
ACL