Probing Script Knowledge from Pre-Trained Models

Zijia Jin, Xingyu Zhang, Mo Yu, Lifu Huang


Abstract
Adversarial attack of structured prediction models faces various challenges such as the difficulty of perturbing discrete words, the sentence quality issue, and the sensitivity of outputs to small perturbations. In this work, we introduce SHARP, a new attack method that formulates the black-box adversarial attack as a search-based optimization problem with a specially designed objective function considering sentence fluency, meaning preservation and attacking effectiveness. Additionally, three different searching strategies are analyzed and compared, , Beam Search, Metropolis-Hastings Sampling, and Hybrid Search. We demonstrate the effectiveness of our attacking strategies on two challenging structured prediction tasks: part-of-speech (POS) tagging and dependency parsing. Through automatic and human evaluations, we show that our method performs a more potent attack compared with pioneer arts. Moreover, the generated adversarial examples can be used to successfully boost the robustness and performance of the victim model via adversarial training.
Anthology ID:
2022.umios-1.10
Volume:
Proceedings of the Workshop on Unimodal and Multimodal Induction of Linguistic Structures (UM-IoS)
Month:
December
Year:
2022
Address:
Abu Dhabi, United Arab Emirates (Hybrid)
Editors:
Wenjuan Han, Zilong Zheng, Zhouhan Lin, Lifeng Jin, Yikang Shen, Yoon Kim, Kewei Tu
Venue:
UM-IoS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
87–93
Language:
URL:
https://aclanthology.org/2022.umios-1.10
DOI:
10.18653/v1/2022.umios-1.10
Bibkey:
Cite (ACL):
Zijia Jin, Xingyu Zhang, Mo Yu, and Lifu Huang. 2022. Probing Script Knowledge from Pre-Trained Models. In Proceedings of the Workshop on Unimodal and Multimodal Induction of Linguistic Structures (UM-IoS), pages 87–93, Abu Dhabi, United Arab Emirates (Hybrid). Association for Computational Linguistics.
Cite (Informal):
Probing Script Knowledge from Pre-Trained Models (Jin et al., UM-IoS 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.umios-1.10.pdf