Enhancing Chat Language Models by Scaling High-quality Instructional Conversations

by   Ning Ding, et al.

Fine-tuning on instruction data has been widely validated as an effective practice for implementing chat language models like ChatGPT. Scaling the diversity and quality of such data, although straightforward, stands a great chance of leading to improved performance. This paper aims to improve the upper bound of open-source models further. We first provide a systematically designed, diverse, informative, large-scale dataset of instructional conversations, UltraChat, which does not involve human queries. Our objective is to capture the breadth of interactions that a human might have with an AI assistant and employs a comprehensive framework to generate multi-turn conversation iteratively. UltraChat contains 1.5 million high-quality multi-turn dialogues and covers a wide range of topics and instructions. Our statistical analysis of UltraChat reveals its superiority in various key metrics, including scale, average length, diversity, coherence, etc., solidifying its position as a leading open-source dataset. Building upon UltraChat, we fine-tune a LLaMA model to create a powerful conversational model, UltraLLaMA. Our evaluations indicate that UltraLLaMA consistently outperforms other open-source models, including Vicuna, the previously recognized state-of-the-art open-source model. The dataset and the model will be publicly released[<https://github.com/thunlp/UltraChat>].


#InsTag: Instruction Tagging for Analyzing Supervised Fine-tuning of Large Language Models

Foundation language models obtain the instruction-following ability thro...

Baize: An Open-Source Chat Model with Parameter-Efficient Tuning on Self-Chat Data

Chat models, such as ChatGPT, have shown impressive capabilities and hav...

MemoChat: Tuning LLMs to Use Memos for Consistent Long-Range Open-Domain Conversation

We propose MemoChat, a pipeline for refining instructions that enables l...

Large Language Model as a User Simulator

The unparalleled performance of closed-sourced ChatGPT has sparked effor...

RefGPT: Reference -> Truthful Customized Dialogues Generation by GPTs and for GPTs

General chat models, like ChatGPT, have attained impressive capability t...

OpenAssistant Conversations – Democratizing Large Language Model Alignment

Aligning large language models (LLMs) with human preferences has proven ...

Please sign up or login with your details

Forgot password? Click here to reset