2023
pdf
bib
abs
Barriers and enabling factors for error analysis in NLG research
Emiel van Miltenburg
|
Miruna Clinciu
|
Ondřej Dušek
|
Dimitra Gkatzia
|
Stephanie Inglis
|
Leo Leppänen
|
Saad Mahamood
|
Stephanie Schoch
|
Craig Thomson
|
Luou Wen
Northern European Journal of Language Technology, Volume 9
Earlier research has shown that few studies in Natural Language Generation (NLG) evaluate their system outputs using an error analysis, despite known limitations of automatic evaluation metrics and human ratings. This position paper takes the stance that error analyses should be encouraged, and discusses several ways to do so. This paper is based on our shared experience as authors as well as a survey we distributed as a means of public consultation. We provide an overview of existing barriers to carrying out error analyses, and propose changes to improve error reporting in the NLG literature.
2021
pdf
bib
abs
Underreporting of errors in NLG output, and what to do about it
Emiel van Miltenburg
|
Miruna Clinciu
|
Ondřej Dušek
|
Dimitra Gkatzia
|
Stephanie Inglis
|
Leo Leppänen
|
Saad Mahamood
|
Emma Manning
|
Stephanie Schoch
|
Craig Thomson
|
Luou Wen
Proceedings of the 14th International Conference on Natural Language Generation
We observe a severe under-reporting of the different kinds of errors that Natural Language Generation systems make. This is a problem, because mistakes are an important indicator of where systems should still be improved. If authors only report overall performance metrics, the research community is left in the dark about the specific weaknesses that are exhibited by ‘state-of-the-art’ research. Next to quantifying the extent of error under-reporting, this position paper provides recommendations for error identification, analysis and reporting.
2017
pdf
bib
abs
Textually Summarising Incomplete Data
Stephanie Inglis
|
Ehud Reiter
|
Somayajulu Sripada
Proceedings of the 10th International Conference on Natural Language Generation
Many data-to-text NLG systems work with data sets which are incomplete, ie some of the data is missing. We have worked with data journalists to understand how they describe incomplete data, and are building NLG algorithms based on these insights. A pilot evaluation showed mixed results, and highlighted several areas where we need to improve our system.
2015
pdf
bib
Summarising Unreliable Data
Stephanie Inglis
Proceedings of the 15th European Workshop on Natural Language Generation (ENLG)