Jangho Kim
2022
Detection of Adversarial Examples in Text Classification: Benchmark and Baseline via Robust Density Estimation
KiYoon Yoo
|
Jangho Kim
|
Jiho Jang
|
Nojun Kwak
Findings of the Association for Computational Linguistics: ACL 2022
Word-level adversarial attacks have shown success in NLP models, drastically decreasing the performance of transformer-based models in recent years. As a countermeasure, adversarial defense has been explored, but relatively few efforts have been made to detect adversarial examples. However, detecting adversarial examples may be crucial for automated tasks (e.g. review sentiment analysis) that wish to amass information about a certain population and additionally be a step towards a robust defense system. To this end, we release a dataset for four popular attack methods on four datasets and four models to encourage further research in this field. Along with it, we propose a competitive baseline based on density estimation that has the highest auc on 29 out of 30 dataset-attack-model combinations. The source code is released (https://github.com/bangawayoo/adversarial-examples-in-text-classification).
Search