Gita Shojaee
2023
RobustQA: A Framework for Adversarial Text Generation Analysis on Question Answering Systems
Yasaman Boreshban
|
Seyed Morteza Mirbostani
|
Seyedeh Fatemeh Ahmadi
|
Gita Shojaee
|
Fatemeh Kamani
|
Gholamreza Ghassem-Sani
|
Seyed Abolghasem Mirroshandel
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: System Demonstrations
Question answering (QA) systems have reached human-level accuracy; however, these systems are not robust enough and are vulnerable to adversarial examples. Recently, adversarial attacks have been widely investigated in text classification. However, there have been few research efforts on this topic in QA. In this article, we have modified the attack algorithms widely used in text classification to fit those algorithms for QA systems. We have evaluated the impact of various attack methods on QA systems at character, word, and sentence levels. Furthermore, we have developed a new framework, named RobustQA, as the first open-source toolkit for investigating textual adversarial attacks in QA systems. RobustQA consists of seven modules: Tokenizer, Victim Model, Goals, Metrics, Attacker, Attack Selector, and Evaluator. It currently supports six different attack algorithms. Furthermore, the framework simplifies the development of new attack algorithms in QA. The source code and documentation of RobustQA are available at https://github.com/mirbostani/RobustQA.
Search