%0 Conference Proceedings %T Chasing the Tail with Domain Generalization: A Case Study on Frequency-Enriched Datasets %A Kumar, Manoj %A Rumshisky, Anna %A Gupta, Rahul %Y He, Yulan %Y Ji, Heng %Y Li, Sujian %Y Liu, Yang %Y Chang, Chua-Hui %S Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 1: Long Papers) %D 2022 %8 November %I Association for Computational Linguistics %C Online only %F kumar-etal-2022-chasing %X Natural language understanding (NLU) tasks are typically defined by creating an annotated dataset in which each utterance is encountered once. Such data does not resemble real-world natural language interactions in which certain utterances are encountered frequently, others rarely. For deployed NLU systems this is a vital problem, since the underlying machine learning (ML) models are often fine-tuned on typical NLU data, and then applied to real-world data with a very different distribution. Such systems need to maintain interpretation consistency for both high-frequency utterances and low-frequency utterances. We propose an alternative strategy that explicitly uses utterance frequency in training data to learn models that are more robust to unknown distributions. We present a methodology to simulate utterance usage in two public NLU corpora and create new corpora with head, body and tail segments. We evaluate several methods for joint intent classification and named entity recognition (IC-NER), and use two domain generalization approaches that we adapt to NER. The proposed approaches demonstrate upto 7.02% relative improvement in semantic accuracy over baselines on the tail data. We provide insights as to why the proposed approaches work and show that the reasons for observed improvements do not align with those reported in previous work. %U https://aclanthology.org/2022.aacl-main.1 %P 1-11