For Generated Text, Is NLI-Neutral Text the Best Text?

Michail Mersinias, Kyle Mahowald


Abstract
We explore incorporating natural language inference (NLI) into the text generative pipeline by using a pre-trained NLI model to assess whether a generated sentence entails, contradicts, or is neutral to the prompt and preceding text. First, we show that the NLI task is predictive of generation errors made by GPT-3. We use these results to develop an NLI-informed generation procedure for GPT-J. Then, we evaluate these generations by obtaining human annotations on error types and overall quality. We find that an NLI strategy of maximizing entailment improves text generation when the nucleus sampling randomness parameter value is high, while one which maximizes contradiction is in fact productive when the parameter value is low. Overall, though, we demonstrate that an NLI strategy of maximizing the neutral class provides the highest quality of generated text (significantly better than the vanilla generations), regardless of parameter value.
Anthology ID:
2023.findings-emnlp.170
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2023
Month:
December
Year:
2023
Address:
Singapore
Editors:
Houda Bouamor, Juan Pino, Kalika Bali
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2596–2602
Language:
URL:
https://aclanthology.org/2023.findings-emnlp.170
DOI:
10.18653/v1/2023.findings-emnlp.170
Bibkey:
Cite (ACL):
Michail Mersinias and Kyle Mahowald. 2023. For Generated Text, Is NLI-Neutral Text the Best Text?. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 2596–2602, Singapore. Association for Computational Linguistics.
Cite (Informal):
For Generated Text, Is NLI-Neutral Text the Best Text? (Mersinias & Mahowald, Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-emnlp.170.pdf