Differential Privacy in Natural Language Processing The Story So Far

Oleksandra Klymenko, Stephen Meisenbacher, Florian Matthes


Abstract
As the tide of Big Data continues to influence the landscape of Natural Language Processing (NLP), the utilization of modern NLP methods has grounded itself in this data, in order to tackle a variety of text-based tasks. These methods without a doubt can include private or otherwise personally identifiable information. As such, the question of privacy in NLP has gained fervor in recent years, coinciding with the development of new Privacy- Enhancing Technologies (PETs). Among these PETs, Differential Privacy boasts several desirable qualities in the conversation surrounding data privacy. Naturally, the question becomes whether Differential Privacy is applicable in the largely unstructured realm of NLP. This topic has sparked novel research, which is unified in one basic goal how can one adapt Differential Privacy to NLP methods? This paper aims to summarize the vulnerabilities addressed by Differential Privacy, the current thinking, and above all, the crucial next steps that must be considered.
Anthology ID:
2022.privatenlp-1.1
Volume:
Proceedings of the Fourth Workshop on Privacy in Natural Language Processing
Month:
July
Year:
2022
Address:
Seattle, United States
Venue:
PrivateNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1–11
Language:
URL:
https://aclanthology.org/2022.privatenlp-1.1
DOI:
10.18653/v1/2022.privatenlp-1.1
Bibkey:
Cite (ACL):
Oleksandra Klymenko, Stephen Meisenbacher, and Florian Matthes. 2022. Differential Privacy in Natural Language Processing The Story So Far. In Proceedings of the Fourth Workshop on Privacy in Natural Language Processing, pages 1–11, Seattle, United States. Association for Computational Linguistics.
Cite (Informal):
Differential Privacy in Natural Language Processing The Story So Far (Klymenko et al., PrivateNLP 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.privatenlp-1.1.pdf