Reconstruction Probing

Najoung Kim, Jatin Khilnani, Alex Warstadt, Abdelrahim Qaddoumi


Abstract
We propose reconstruction probing, a new analysis method for contextualized representations based on reconstruction probabilities in masked language models (MLMs). This method relies on comparing the reconstruction probabilities of tokens in a given sequence when conditioned on the representation of a single token that has been fully contextualized and when conditioned on only the decontextualized lexical prior of the model. This comparison can be understood as quantifying the contribution of contextualization towards reconstruction—the difference in the reconstruction probabilities can only be attributed to the representational change of the single token induced by contextualization. We apply this analysis to three MLMs and find that contextualization boosts reconstructability of tokens that are close to the token being reconstructed in terms of linear and syntactic distance. Furthermore, we extend our analysis to finer-grained decomposition of contextualized representations, and we find that these boosts are largely attributable to static and positional embeddings at the input layer.
Anthology ID:
2023.findings-acl.523
Volume:
Findings of the Association for Computational Linguistics: ACL 2023
Month:
July
Year:
2023
Address:
Toronto, Canada
Editors:
Anna Rogers, Jordan Boyd-Graber, Naoaki Okazaki
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
8240–8255
Language:
URL:
https://aclanthology.org/2023.findings-acl.523
DOI:
10.18653/v1/2023.findings-acl.523
Bibkey:
Cite (ACL):
Najoung Kim, Jatin Khilnani, Alex Warstadt, and Abdelrahim Qaddoumi. 2023. Reconstruction Probing. In Findings of the Association for Computational Linguistics: ACL 2023, pages 8240–8255, Toronto, Canada. Association for Computational Linguistics.
Cite (Informal):
Reconstruction Probing (Kim et al., Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-acl.523.pdf