Gold Doesn’t Always Glitter: Spectral Removal of Linear and Nonlinear Guarded Attribute Information

Shun Shao, Yftah Ziser, Shay B. Cohen


Abstract
We describe a simple and effective method (Spectral Attribute removaL; SAL) to remove private or guarded information from neural representations. Our method uses matrix decomposition to project the input representations into directions with reduced covariance with the guarded information rather than maximal covariance as factorization methods normally use. We begin with linear information removal and proceed to generalize our algorithm to the case of nonlinear information removal using kernels. Our experiments demonstrate that our algorithm retains better main task performance after removing the guarded information compared to previous work. In addition, our experiments demonstrate that we need a relatively small amount of guarded attribute data to remove information about these attributes, which lowers the exposure to sensitive data and is more suitable for low-resource scenarios.
Anthology ID:
2023.eacl-main.118
Original:
2023.eacl-main.118v1
Version 2:
2023.eacl-main.118v2
Volume:
Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics
Month:
May
Year:
2023
Address:
Dubrovnik, Croatia
Editors:
Andreas Vlachos, Isabelle Augenstein
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1611–1622
Language:
URL:
https://aclanthology.org/2023.eacl-main.118
DOI:
10.18653/v1/2023.eacl-main.118
Bibkey:
Cite (ACL):
Shun Shao, Yftah Ziser, and Shay B. Cohen. 2023. Gold Doesn’t Always Glitter: Spectral Removal of Linear and Nonlinear Guarded Attribute Information. In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, pages 1611–1622, Dubrovnik, Croatia. Association for Computational Linguistics.
Cite (Informal):
Gold Doesn’t Always Glitter: Spectral Removal of Linear and Nonlinear Guarded Attribute Information (Shao et al., EACL 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.eacl-main.118.pdf
Video:
 https://aclanthology.org/2023.eacl-main.118.mp4