The Unreasonable Effectiveness of Easy Training Data for Hard Tasks

Peter Hase, Mohit Bansal, Peter Clark, Sarah Wiegreffe


Abstract
How can we train models to perform well on hard test data when hard training data is by definition difficult to label correctly? This question has been termed the scalable oversight problem and has drawn increasing attention as language models have continually improved. In this paper, we present the surprising conclusion that current pretrained language models often generalize relatively well from easy to hard data, even performing as well as oracle models finetuned on hard data. We demonstrate this kind of easy-to-hard generalization using simple finetuning methods like in-context learning, linear classifier heads, and QLoRA for seven different measures of datapoint hardness, including six empirically diverse human hardness measures (like grade level) and one model-based measure (loss-based). Furthermore, we show that even if one cares most about model performance on hard data, it can be better to collect easy data rather than hard data for finetuning, since hard data is generally noisier and costlier to collect. Our experiments use open models up to 70b in size and four publicly available question-answering datasets with questions ranging in difficulty from 3rd grade science questions to college level STEM questions and general-knowledge trivia. We conclude that easy-to-hard generalization in LMs is surprisingly strong for the tasks studied.
Anthology ID:
2024.acl-long.378
Volume:
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
August
Year:
2024
Address:
Bangkok, Thailand
Editors:
Lun-Wei Ku, Andre Martins, Vivek Srikumar
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
7002–7024
Language:
URL:
https://aclanthology.org/2024.acl-long.378
DOI:
Bibkey:
Cite (ACL):
Peter Hase, Mohit Bansal, Peter Clark, and Sarah Wiegreffe. 2024. The Unreasonable Effectiveness of Easy Training Data for Hard Tasks. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 7002–7024, Bangkok, Thailand. Association for Computational Linguistics.
Cite (Informal):
The Unreasonable Effectiveness of Easy Training Data for Hard Tasks (Hase et al., ACL 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.acl-long.378.pdf