%0 Conference Proceedings %T Identifying the Limits of Cross-Domain Knowledge Transfer for Pretrained Models %A Wu, Zhengxuan %A F. Liu, Nelson %A Potts, Christopher %Y Gella, Spandana %Y He, He %Y Majumder, Bodhisattwa Prasad %Y Can, Burcu %Y Giunchiglia, Eleonora %Y Cahyawijaya, Samuel %Y Min, Sewon %Y Mozes, Maximilian %Y Li, Xiang Lorraine %Y Augenstein, Isabelle %Y Rogers, Anna %Y Cho, Kyunghyun %Y Grefenstette, Edward %Y Rimell, Laura %Y Dyer, Chris %S Proceedings of the 7th Workshop on Representation Learning for NLP %D 2022 %8 May %I Association for Computational Linguistics %C Dublin, Ireland %F wu-etal-2022-identifying %X There is growing evidence that pretrained language models improve task-specific fine-tuning even where the task examples are radically different from those seen in training. We study an extreme case of transfer learning by providing a systematic exploration of how much transfer occurs when models are denied any information about word identity via random scrambling. In four classification tasks and two sequence labeling tasks, we evaluate LSTMs using GloVe embeddings, BERT, and baseline models. Among these models, we find that only BERT shows high rates of transfer into our scrambled domains, and for classification but not sequence labeling tasks. Our analyses seek to explain why transfer succeeds for some tasks but not others, to isolate the separate contributions of pretraining versus fine-tuning, to show that the fine-tuning process is not merely learning to unscramble the scrambled inputs, and to quantify the role of word frequency. Furthermore, our results suggest that current benchmarks may overestimate the degree to which current models actually understand language. %R 10.18653/v1/2022.repl4nlp-1.11 %U https://aclanthology.org/2022.repl4nlp-1.11 %U https://doi.org/10.18653/v1/2022.repl4nlp-1.11 %P 100-110