RoFT: A Tool for Evaluating Human Detection of Machine-Generated Text

Liam Dugan, Daphne Ippolito, Arun Kirubarajan, Chris Callison-Burch


Abstract
In recent years, large neural networks for natural language generation (NLG) have made leaps and bounds in their ability to generate fluent text. However, the tasks of evaluating quality differences between NLG systems and understanding how humans perceive the generated text remain both crucial and difficult. In this system demonstration, we present Real or Fake Text (RoFT), a website that tackles both of these challenges by inviting users to try their hand at detecting machine-generated text in a variety of domains. We introduce a novel evaluation task based on detecting the boundary at which a text passage that starts off human-written transitions to being machine-generated. We show preliminary results of using RoFT to evaluate detection of machine-generated news articles.
Anthology ID:
2020.emnlp-demos.25
Volume:
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations
Month:
October
Year:
2020
Address:
Online
Editors:
Qun Liu, David Schlangen
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
189–196
Language:
URL:
https://aclanthology.org/2020.emnlp-demos.25
DOI:
10.18653/v1/2020.emnlp-demos.25
Bibkey:
Cite (ACL):
Liam Dugan, Daphne Ippolito, Arun Kirubarajan, and Chris Callison-Burch. 2020. RoFT: A Tool for Evaluating Human Detection of Machine-Generated Text. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 189–196, Online. Association for Computational Linguistics.
Cite (Informal):
RoFT: A Tool for Evaluating Human Detection of Machine-Generated Text (Dugan et al., EMNLP 2020)
Copy Citation:
PDF:
https://aclanthology.org/2020.emnlp-demos.25.pdf
Code
 kirubarajan/roft +  additional community code