Unlearn What You Want to Forget: Efficient Unlearning for LLMs

Jiaao Chen, Diyi Yang


Abstract
Large language models (LLMs) have achieved significant progress from pre-training on and memorizing a wide range of textual data, however, this process might suffer from privacy issues and violations of data protection regulations. As a result, the ability to easily remove data related to individual users from such models while not deteriorating their predictive quality after the removal becomes increasingly important. To address these issues, in this work, we propose an efficient unlearning framework that could efficiently update LLMs without having to retrain the whole model after data removals, by introducing lightweight unlearning layers learned with a selective teacher-student objective into the transformers. In addition, we introduce a fusion mechanism to effectively combine different unlearning layers that learns to forget different sets of data to handle a sequence of forgetting operations. Experiments on classification and generation tasks demonstrate the effectiveness of our proposed methods compared to the state-of-the-art baselines.
Anthology ID:
2023.emnlp-main.738
Volume:
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing
Month:
December
Year:
2023
Address:
Singapore
Editors:
Houda Bouamor, Juan Pino, Kalika Bali
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
12041–12052
Language:
URL:
https://aclanthology.org/2023.emnlp-main.738
DOI:
10.18653/v1/2023.emnlp-main.738
Bibkey:
Cite (ACL):
Jiaao Chen and Diyi Yang. 2023. Unlearn What You Want to Forget: Efficient Unlearning for LLMs. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 12041–12052, Singapore. Association for Computational Linguistics.
Cite (Informal):
Unlearn What You Want to Forget: Efficient Unlearning for LLMs (Chen & Yang, EMNLP 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.emnlp-main.738.pdf
Video:
 https://aclanthology.org/2023.emnlp-main.738.mp4