%0 Conference Proceedings %T Towards Robust NLG Bias Evaluation with Syntactically-diverse Prompts %A Aggarwal, Arshiya %A Sun, Jiao %A Peng, Nanyun %Y Goldberg, Yoav %Y Kozareva, Zornitsa %Y Zhang, Yue %S Findings of the Association for Computational Linguistics: EMNLP 2022 %D 2022 %8 December %I Association for Computational Linguistics %C Abu Dhabi, United Arab Emirates %F aggarwal-etal-2022-towards %X We present a robust methodology for evaluating biases in natural language generation(NLG) systems. Previous works use fixed hand-crafted prefix templates with mentions of various demographic groups to prompt models to generate continuations for bias analysis. These fixed prefix templates could themselves be specific in terms of styles or linguistic structures, which may lead to unreliable fairness conclusions that are not representative of the general trends from tone varying prompts. To study this problem, we paraphrase the prompts with different syntactic structures and use these to evaluate demographic bias in NLG systems. Our results suggest similar overall bias trends but some syntactic structures lead to contradictory conclusions compared to past works. We show that our methodology is more robust and that some syntactic structures prompt more toxic content while others could prompt less biased generation. This suggests the importance of not relying on a fixed syntactic structure and using tone-invariant prompts. Introducing syntactically-diverse prompts can achieve more robust NLG (bias) evaluation. %R 10.18653/v1/2022.findings-emnlp.445 %U https://aclanthology.org/2022.findings-emnlp.445 %U https://doi.org/10.18653/v1/2022.findings-emnlp.445 %P 6022-6032