Margaret E. Roberts
2022
Causal Inference in Natural Language Processing: Estimation, Prediction, Interpretation and Beyond
Amir Feder
|
Katherine A. Keith
|
Emaad Manzoor
|
Reid Pryzant
|
Dhanya Sridhar
|
Zach Wood-Doughty
|
Jacob Eisenstein
|
Justin Grimmer
|
Roi Reichart
|
Margaret E. Roberts
|
Brandon M. Stewart
|
Victor Veitch
|
Diyi Yang
Transactions of the Association for Computational Linguistics, Volume 10
A fundamental goal of scientific research is to learn about causal relationships. However, despite its critical role in the life and social sciences, causality has not had the same importance in Natural Language Processing (NLP), which has traditionally placed more emphasis on predictive tasks. This distinction is beginning to fade, with an emerging area of interdisciplinary research at the convergence of causal inference and language processing. Still, research on causality in NLP remains scattered across domains without unified definitions, benchmark datasets and clear articulations of the challenges and opportunities in the application of causal inference to the textual domain, with its unique properties. In this survey, we consolidate research across academic areas and situate it in the broader NLP landscape. We introduce the statistical challenge of estimating causal effects with text, encompassing settings where text is used as an outcome, treatment, or to address confounding. In addition, we explore potential uses of causal inference to improve the robustness, fairness, and interpretability of NLP models. We thus provide a unified overview of causal inference for the NLP community.1
2015
TopicCheck: Interactive Alignment for Assessing Topic Model Stability
Jason Chuang
|
Margaret E. Roberts
|
Brandon M. Stewart
|
Rebecca Weiss
|
Dustin Tingley
|
Justin Grimmer
|
Jeffrey Heer
Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies