%0 Conference Proceedings %T Automatic Discrimination of Human and Neural Machine Translation: A Study with Multiple Pre-Trained Models and Longer Context %A van der Werff, Tobias %A van Noord, Rik %A Toral, Antonio %Y Moniz, Helena %Y Macken, Lieve %Y Rufener, Andrew %Y Barrault, Loïc %Y Costa-jussà, Marta R. %Y Declercq, Christophe %Y Koponen, Maarit %Y Kemp, Ellie %Y Pilos, Spyridon %Y Forcada, Mikel L. %Y Scarton, Carolina %Y Van den Bogaert, Joachim %Y Daems, Joke %Y Tezcan, Arda %Y Vanroy, Bram %Y Fonteyne, Margot %S Proceedings of the 23rd Annual Conference of the European Association for Machine Translation %D 2022 %8 June %I European Association for Machine Translation %C Ghent, Belgium %F van-der-werff-etal-2022-automatic %X We address the task of automatically distinguishing between human-translated (HT) and machine translated (MT) texts. Following recent work, we fine-tune pre-trained language models (LMs) to perform this task. Our work differs in that we use state-of-the-art pre-trained LMs, as well as the test sets of the WMT news shared tasks as training data, to ensure the sentences were not seen during training of the MT system itself. Moreover, we analyse performance for a number of different experimental setups, such as adding translationese data, going beyond the sentence-level and normalizing punctuation. We show that (i) choosing a state-of-the-art LM can make quite a difference: our best baseline system (DeBERTa) outperforms both BERT and RoBERTa by over 3% accuracy, (ii) adding translationese data is only beneficial if there is not much data available, (iii) considerable improvements can be obtained by classifying at the document-level and (iv) normalizing punctuation and thus avoiding (some) shortcuts has no impact on model performance. %U https://aclanthology.org/2022.eamt-1.19 %P 161-170