Bokai Xu
2023
Enhancing Chat Language Models by Scaling High-quality Instructional Conversations
Ning Ding
|
Yulin Chen
|
Bokai Xu
|
Yujia Qin
|
Shengding Hu
|
Zhiyuan Liu
|
Maosong Sun
|
Bowen Zhou
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing
Fine-tuning on instruction data has been widely validated as an effective practice for implementing chat language models like ChatGPT. Scaling the diversity and quality of such data, although straightforward, stands a great chance of leading to improved performance. This paper aims to push the upper bound of open-source models further. We first provide a systematically designed, diverse, informative, large-scale dataset of instructional conversations, UltraChat, which does not involve human queries. Our objective is to capture the breadth of interactions between a human user and an AI assistant and employs a comprehensive framework to generate multi-turn conversation iteratively. UltraChat contains 1.5 million high-quality multi-turn dialogues and covers a wide range of topics and instructions. Our statistical analysis of UltraChat reveals its superiority in various key metrics, including scale, average length, diversity, coherence, etc., solidifying its position as a leading open-source dataset. Building upon UltraChat, we fine-tune a LLaMA model to create a powerful conversational model, UltraLM. Our evaluations indicate that UltraLM consistently outperforms other open-source models, including WizardLM and Vicuna, the previously recognized state-of-the-art open-source models.
Search
Co-authors
- Ning Ding 1
- Yulin Chen 1
- Yujia Qin 1
- Shengding Hu 1
- Zhiyuan Liu 1
- show all...