Bias Identification and Attribution in NLP Models With Regression and Effect Sizes

Erenay Dayanik, Ngoc Thang Vu, Sebastian Padó


Abstract
In recent years, there has been an increasing awareness that many NLP systems incorporate biases of various types (e.g., regarding gender or race) which can have significant negative consequences. At the same time, the techniques used to statistically analyze such biases are still relatively simple. Typically, studies test for the presence of a significant difference between two levels of a single bias variable (e.g., male vs. female) without attention to potential confounders, and do not quantify the importance of the bias variable. This article proposes to analyze bias in the output of NLP systems using multivariate regression models. They provide a robust and more informative alternative which (a) generalizes to multiple bias variables, (b) can take covariates into account, (c) can be combined with measures of effect size to quantify the size of bias. Jointly, these effects contribute to a more robust statistical analysis of bias that can be used to diagnose system behavior and extract informative examples. We demonstrate the benefits of our method by analyzing a range of current NLP models on one regression and one classification tasks (emotion intensity prediction and coreference resolution, respectively).
Anthology ID:
2022.nejlt-1.4
Volume:
Northern European Journal of Language Technology, Volume 8
Month:
Year:
2022
Address:
Copenhagen, Denmark
Editor:
Leon Derczynski
Venue:
NEJLT
SIG:
Publisher:
Northern European Association of Language Technology
Note:
Pages:
Language:
URL:
https://aclanthology.org/2022.nejlt-1.4
DOI:
https://doi.org/10.3384/nejlt.2000-1533.2022.3505
Bibkey:
Cite (ACL):
Erenay Dayanik, Ngoc Thang Vu, and Sebastian Padó. 2022. Bias Identification and Attribution in NLP Models With Regression and Effect Sizes. In Northern European Journal of Language Technology, Volume 8, Copenhagen, Denmark. Northern European Association of Language Technology.
Cite (Informal):
Bias Identification and Attribution in NLP Models With Regression and Effect Sizes (Dayanik et al., NEJLT 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.nejlt-1.4.pdf