Tianjiao Yu
2025
MOCHA: Are Code Language Models Robust Against Multi-Turn Malicious Coding Prompts?
Muntasir Wahed
|
Xiaona Zhou
|
Kiet A. Nguyen
|
Tianjiao Yu
|
Nirav Diwan
|
Gang Wang
|
Dilek Hakkani-Tür
|
Ismini Lourentzou
Findings of the Association for Computational Linguistics: EMNLP 2025
Recent advancements in Large Language Models (LLMs) have significantly enhanced their code generation capabilities. However, their robustness against adversarial misuse, particularly through multi-turn malicious coding prompts, remains underexplored. In this work, we introduce code decomposition attacks, where a malicious coding task is broken down into a series of seemingly benign subtasks across multiple conversational turns to evade safety filters. To facilitate systematic evaluation, we introduce MOCHA, a large-scale benchmark designed to evaluate the robustness of code LLMs against both single-turn and multi-turn malicious prompts. Empirical results across open- and closed-source models reveal persistent vulnerabilities, especially under multi-turn scenarios. Fine-tuning on MOCHA improves rejection rates while preserving coding ability, and importantly, enhances robustness on external adversarial datasets with up to 32.4% increase in rejection rates without any additional supervision.
Search
Fix author
Co-authors
- Nirav Diwan 1
- Dilek Hakkani-Tur 1
- Ismini Lourentzou 1
- Kiet A. Nguyen 1
- Muntasir Wahed 1
- show all...