@inproceedings{kmjec-bojar-2023-team,
title = "Team Iterate @ {A}uto{M}in 2023 - Experiments with Iterative Minuting",
author = "Kmje{\v{c}}, Franti{\v{s}}ek and
Bojar, Ond{\v{r}}ej",
editor = "Mille, Simon",
booktitle = "Proceedings of the 16th International Natural Language Generation Conference: Generation Challenges",
month = sep,
year = "2023",
address = "Prague, Czechia",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.inlg-genchal.16/",
pages = "114--120",
abstract = "This report describes the development of our system for automatic minuting created for the AutoMin 2023 Task A. As a baseline, we utilize a system based on the BART encoder-decoder model paired with a preprocessing pipeline similar to the one introduced by the winning solutions at AutoMin 2021. We then further explore the possibilities for iterative summarization by constructing an iterative minuting dataset from the provided data, finetuning on it and feeding the model previously generated minutes. We also experiment with adding more context by utilizing the Longformer encoder-decoder model and finetuning it on the SAMSum dataset. Our submitted solution is of the baseline approach, since we were unable to match its performance with our iterative variants. With the baseline, we achieve a ROUGE-1 score of 0.368 on the ELITR minuting corpus development set. We finally explore the performance of Vicuna 13B quantized language model for summarization."
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="kmjec-bojar-2023-team">
<titleInfo>
<title>Team Iterate @ AutoMin 2023 - Experiments with Iterative Minuting</title>
</titleInfo>
<name type="personal">
<namePart type="given">František</namePart>
<namePart type="family">Kmječ</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Ondřej</namePart>
<namePart type="family">Bojar</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2023-09</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 16th International Natural Language Generation Conference: Generation Challenges</title>
</titleInfo>
<name type="personal">
<namePart type="given">Simon</namePart>
<namePart type="family">Mille</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Prague, Czechia</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>This report describes the development of our system for automatic minuting created for the AutoMin 2023 Task A. As a baseline, we utilize a system based on the BART encoder-decoder model paired with a preprocessing pipeline similar to the one introduced by the winning solutions at AutoMin 2021. We then further explore the possibilities for iterative summarization by constructing an iterative minuting dataset from the provided data, finetuning on it and feeding the model previously generated minutes. We also experiment with adding more context by utilizing the Longformer encoder-decoder model and finetuning it on the SAMSum dataset. Our submitted solution is of the baseline approach, since we were unable to match its performance with our iterative variants. With the baseline, we achieve a ROUGE-1 score of 0.368 on the ELITR minuting corpus development set. We finally explore the performance of Vicuna 13B quantized language model for summarization.</abstract>
<identifier type="citekey">kmjec-bojar-2023-team</identifier>
<location>
<url>https://aclanthology.org/2023.inlg-genchal.16/</url>
</location>
<part>
<date>2023-09</date>
<extent unit="page">
<start>114</start>
<end>120</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Team Iterate @ AutoMin 2023 - Experiments with Iterative Minuting
%A Kmječ, František
%A Bojar, Ondřej
%Y Mille, Simon
%S Proceedings of the 16th International Natural Language Generation Conference: Generation Challenges
%D 2023
%8 September
%I Association for Computational Linguistics
%C Prague, Czechia
%F kmjec-bojar-2023-team
%X This report describes the development of our system for automatic minuting created for the AutoMin 2023 Task A. As a baseline, we utilize a system based on the BART encoder-decoder model paired with a preprocessing pipeline similar to the one introduced by the winning solutions at AutoMin 2021. We then further explore the possibilities for iterative summarization by constructing an iterative minuting dataset from the provided data, finetuning on it and feeding the model previously generated minutes. We also experiment with adding more context by utilizing the Longformer encoder-decoder model and finetuning it on the SAMSum dataset. Our submitted solution is of the baseline approach, since we were unable to match its performance with our iterative variants. With the baseline, we achieve a ROUGE-1 score of 0.368 on the ELITR minuting corpus development set. We finally explore the performance of Vicuna 13B quantized language model for summarization.
%U https://aclanthology.org/2023.inlg-genchal.16/
%P 114-120
Markdown (Informal)
[Team Iterate @ AutoMin 2023 - Experiments with Iterative Minuting](https://aclanthology.org/2023.inlg-genchal.16/) (Kmječ & Bojar, INLG-SIGDIAL 2023)
ACL