METAPROBE: A Representation- and Task-Agnostic Probe

Yichu Zhou, Vivek Srikumar


Abstract
Probing contextualized representations typically involves comparing task-specific model predictions against ground truth linguistic labels. Although this methodology shows what information can be recovered by a classifier, it does not reveal how a classifier uses the representation to make its decision. To address the latter problem, we ask: Do task-classifiers rely on representation- and task-independent geometric patterns in the embedding space? We explore this question by developing MetaProbe, an approach that uses geometric properties of representations to predict the behavior of task-specific classifiers (i.e., their predictions as opposed to the ground truth). Our experiments reveal the existence of universal geometric patterns across representations that can predict classifier predictions. Consequently, this allows us to posit a geometric explanation for the impressive performance of contextualized representations.
Anthology ID:
2023.blackboxnlp-1.18
Volume:
Proceedings of the 6th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP
Month:
December
Year:
2023
Address:
Singapore
Editors:
Yonatan Belinkov, Sophie Hao, Jaap Jumelet, Najoung Kim, Arya McCarthy, Hosein Mohebbi
Venues:
BlackboxNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
233–249
Language:
URL:
https://aclanthology.org/2023.blackboxnlp-1.18
DOI:
10.18653/v1/2023.blackboxnlp-1.18
Bibkey:
Cite (ACL):
Yichu Zhou and Vivek Srikumar. 2023. METAPROBE: A Representation- and Task-Agnostic Probe. In Proceedings of the 6th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP, pages 233–249, Singapore. Association for Computational Linguistics.
Cite (Informal):
METAPROBE: A Representation- and Task-Agnostic Probe (Zhou & Srikumar, BlackboxNLP-WS 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.blackboxnlp-1.18.pdf