%0 Conference Proceedings %T Towards Identifying Social Bias in Dialog Systems: Framework, Dataset, and Benchmark %A Zhou, Jingyan %A Deng, Jiawen %A Mi, Fei %A Li, Yitong %A Wang, Yasheng %A Huang, Minlie %A Jiang, Xin %A Liu, Qun %A Meng, Helen %Y Goldberg, Yoav %Y Kozareva, Zornitsa %Y Zhang, Yue %S Findings of the Association for Computational Linguistics: EMNLP 2022 %D 2022 %8 December %I Association for Computational Linguistics %C Abu Dhabi, United Arab Emirates %F zhou-etal-2022-towards-identifying %X Among all the safety concerns that hinder the deployment of open-domain dialog systems (e.g., offensive languages, biases, and toxic behaviors), social bias presents an insidious challenge. Addressing this challenge requires rigorous analyses and normative reasoning. In this paper, we focus our investigation on social bias measurement to facilitate the development of unbiased dialog systems. We first propose a novel Dial-Bias Framework for analyzing the social bias in conversations using a holistic method beyond bias lexicons or dichotomous annotations. Leveraging the proposed framework, we further introduce the CDial-Bias Dataset which is, to the best of our knowledge, the first annotated Chinese social bias dialog dataset. We also establish a fine-grained dialog bias measurement benchmark and conduct in-depth ablation studies to shed light on the utility of the detailed annotations in the proposed dataset. Finally, we evaluate representative Chinese generative models with our classifiers to unveil the presence of social bias in these systems. %R 10.18653/v1/2022.findings-emnlp.262 %U https://aclanthology.org/2022.findings-emnlp.262 %U https://doi.org/10.18653/v1/2022.findings-emnlp.262 %P 3576-3591