Home  >  Article  >  Technology peripherals  >  Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

PHPz
PHPzforward
2023-04-14 13:16:11998browse

The dream of having a ChatGPT for everyone is about to come true?

Just now, Microsoft has open sourced a system framework that can add a complete RLHF process to model training - DeepSpeed ​​Chat.

In other words, high-quality ChatGPT-like models of all sizes are now within easy reach!

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

## Project address: https://github.com/microsoft/DeepSpeed

Unlock 100 billion-level ChatGPT with one click and easily save 15 times

As we all know, because OpenAI is not Open, the open source community has successively launched LLaMa and Alpaca in order to allow more people to use ChatGPT-like models. , Vicuna, Databricks-Dolly and other models.

However, due to the lack of an end-to-end RLHF scale system, the current training of ChatGPT-like models is still very difficult. The emergence of DeepSpeed ​​Chat just makes up for this "bug".

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

What’s even brighter is that DeepSpeed ​​Chat has greatly reduced costs.

Previously, expensive multi-GPU setups were beyond the capabilities of many researchers, and even with access to multi-GPU clusters, existing methods cannot afford hundreds of billions of parameter ChatGPT models training.

Now, for only $1,620, you can train an OPT-66B model in 2.1 days through the hybrid engine DeepSpeed-HE.

If you use a multi-node, multi-GPU system, DeepSpeed-HE can spend 320 US dollars to train an OPT-13B model in 1.25 hours, and spend 5120 US dollars to train an OPT-13B model in less than 1.25 hours. Train an OPT-175B model in one day.

Former Meta AI expert Elvis forwarded it excitedly, saying this was a big deal and expressed curiosity about how DeepSpeed ​​Chat compared with ColossalChat.

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

Next, let’s take a look at the effect.

After training by DeepSpeed-Chat, the 1.3 billion parameter version "ChatGPT" performed very well in the question and answer session. Not only can you get the context of the question, but the answers given are also relevant.

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

In multiple rounds of conversations, the performance demonstrated by this 1.3 billion parameter version of "ChatGPT" completely exceeded the inherent impression of this scale. .

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

A piece of code to generate your first ChatGPT

Of course, before you experience it, you need to configure the environment:

git clone https://github.com/microsoft/DeepSpeed.git
cd DeepSpeed
pip install .


git clone https://github.com/microsoft/DeepSpeedExamples.git
cd DeepSpeedExamples/applications/DeepSpeed-Chat/
pip install -r requirements.txt

A cup of coffee, after training the 1.3 billion parameter version of ChatGPT

If you only have about 1-2 hours of coffee or lunch break, You can also try to use DeepSpeed-Chat to train a "little toy".

The team has specially prepared a training example for the 1.3B model, which can be tested on consumer-grade GPUs. The best part is, when you come back from your lunch break, everything is ready.

python train.py --actor-model facebook/opt-1.3b --reward-model facebook/opt-350m --num-gpus 1

Consumer-grade NVIDIA A6000 GPU with 48GB of video memory:

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

一个GPU Node,半天搞定130亿参数

如果你只有半天的时间,以及一台服务器节点,则可以通过预训练的OPT-13B作为actor模型,OPT-350M作为reward模型,来生成一个130亿参数的类ChatGPT模型:

python train.py --actor-model facebook/opt-13b --reward-model facebook/opt-350m --num-gpus 8

单DGX节点,搭载了8个NVIDIA A100-40G GPU:

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

超省钱云方案,训练660亿参数模型

如果你可以使用多节点集群或云资源,并希望训练一个更大、更高质量的模型。那么只需基于下面这行代码,输入你想要的模型大小(如66B)和GPU数量(如64):

python train.py --actor-model facebook/opt-66b --reward-model facebook/opt-350m --num-gpus 64

8个DGX节点,每个节点配备8个NVIDIA A100-80G GPU:

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

具体来说,针对不同规模的模型和硬件配置,DeepSpeed-RLHF系统所需的时间和成本如下:

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

DeepSpeed Chat是个啥?

DeepSpeed Chat是一种通用系统框架,能够实现类似ChatGPT模型的端到端RLHF训练,从而帮助我们生成自己的高质量类ChatGPT模型。

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

DeepSpeed Chat具有以下三大核心功能:

1. 简化ChatGPT类型模型的训练和强化推理体验

开发者只需一个脚本,就能实现多个训练步骤,并且在完成后还可以利用推理API进行对话式交互测试。

2. DeepSpeed-RLHF模块

DeepSpeed-RLHF复刻了InstructGPT论文中的训练模式,并提供了数据抽象和混合功能,支持开发者使用多个不同来源的数据源进行训练。

3. DeepSpeed-RLHF系统

团队将DeepSpeed的训练(training engine)和推理能力(inference engine) 整合成了一个统一的混合引擎(DeepSpeed Hybrid Engine or DeepSpeed-HE)中,用于RLHF训练。由于,DeepSpeed-HE能够无缝地在推理和训练模式之间切换,因此可以利用来自DeepSpeed-Inference的各种优化。

DeepSpeed-RLHF系统在大规模训练中具有无与伦比的效率,使复杂的RLHF训练变得快速、经济并且易于大规模推广:

  • 高效且经济:

DeepSpeed-HE比现有系统快15倍以上,使RLHF训练快速且经济实惠。

例如,DeepSpeed-HE在Azure云上只需9小时即可训练一个OPT-13B模型,只需18小时即可训练一个OPT-30B模型。这两种训练分别花费不到300美元和600美元。

  • Excellent scalability:

DeepSpeed-HE can support training models with hundreds of billions of parameters and run on multi-node multi-GPU systems Demonstrates excellent scalability.

Therefore, even a model with 13 billion parameters can be trained in only 1.25 hours. For a model with 175 billion parameters, training using DeepSpeed-HE only takes less than a day.

  • Popularizing RLHF training:

Only a single GPU, DeepSpeed-HE can support training for more than A model with 13 billion parameters. This enables data scientists and researchers who do not have access to multi-GPU systems to easily create not only lightweight RLHF models, but also large and powerful models to cope with different usage scenarios.

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

Complete RLHF training process

In order to provide a seamless training experience, the researcher followed InstructGPT and used it in DeepSpeed-Chat It contains a complete end-to-end training process.

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

##DeepSpeed-Chat’s RLHF training process diagram, including some optional functions

The process consists of three main steps:

    Step 1:

Supervised fine-tuning (SFT), using Featured human answers to fine-tune pre-trained language models to respond to various queries.

    Step 2:

Reward model fine-tuning, trained with a dataset containing human ratings of multiple answers to the same query An independent (usually smaller than SFT) reward model (RW).

    Step 3:

RLHF training. In this step, the SFT model is trained from the RW model by using the approximate policy optimization (PPO) algorithm. The reward feedback is further fine-tuned.

In step 3, the researcher also provides two additional functions to help improve the quality of the model:

-Exponential Moving Average (EMA) collection, you can select an EMA-based checkpoint for final evaluation.

- Hybrid training, mixing the pre-training objective (i.e. next word prediction) with the PPO objective to prevent performance regression on public benchmarks (such as SQuAD2.0).

EMA and hybrid training are two training features that are often ignored by other open source frameworks because they do not hinder training.

However, according to InstructGPT, EMA checkpoints tend to provide better response quality than traditional final training models, and hybrid training can help the model maintain pre-training baseline solving capabilities.

Therefore, the researchers provide these functions for users so that they can fully obtain the training experience described in InstructGPT.

In addition to being highly consistent with the InstructGPT paper, the researcher also provides functions that allow developers to use a variety of data resources to train their own RLHF models:

    Data abstraction and hybrid capabilities:

DeepSpeed-Chat is equipped with (1) an abstract dataset layer to unify the format of different datasets; and (2) data splitting Split/Blend functionality, whereby multiple datasets are blended appropriately and then split in 3 training stages.

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

DeepSpeed ​​Hybrid Engine

Indicates steps 1 and 2 of the guided RLHF pipeline, similar to conventional fine-tuning of large models, which are driven by ZeRO-based optimization and flexible parallelism in DeepSpeed ​​training A combination of strategies to achieve scale and speed.

And step 3 of the pipeline is the most complex part in terms of performance impact.

Each iteration needs to effectively handle two stages: a) the inference stage, used for token/experience generation, to generate training input; b) the training stage, updating actors and The weights of the reward model, as well as the interaction and scheduling between them.

It introduces two major difficulties: (1) memory cost, since multiple SFT and RW models need to be run throughout the third stage; (2) speed of the answer generation stage Slower, if not accelerated correctly, will significantly slow down the entire third stage.

In addition, the two important functions added by the researchers in the third stage-exponential moving average (EMA) collection and hybrid training will incur additional memory and training costs.

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

#To address these challenges, researchers have combined all system capabilities of DeepSpeed ​​training and inference into a unified infrastructure , that is, Hybrid Engine.

It utilizes the original DeepSpeed ​​engine for fast training mode, while effortlessly applying the DeepSpeed ​​inference engine for generation/evaluation mode, providing a better solution for the third stage of RLHF training. Fast training system.

As shown in the figure below, the transition between the DeepSpeed ​​training and inference engines is seamless: by enabling the typical eval and train modes for the actor model, when running the inference and training processes , DeepSpeed ​​has chosen different optimizations to run the model faster and improve the throughput of the overall system.

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

##DeepSpeed ​​hybrid engine design to accelerate the most time-consuming part of the RLHF process

During the inference execution process of the experience generation phase of RLHF training, the DeepSpeed ​​hybrid engine uses a lightweight memory management system to handle KV cache and intermediate results, while using highly optimized inference CUDA cores and tensor parallel computing , compared with existing solutions, the throughput (number of tokens per second) has been greatly improved.

During training, the hybrid engine enables memory optimization technologies such as DeepSpeed’s ZeRO family of technologies and Low-Order Adaptation (LoRA).

The way researchers design and implement optimization of these systems is to make them compatible with each other and can be combined together to provide the highest training efficiency under a unified hybrid engine.

The hybrid engine can seamlessly change model partitioning during training and inference to support tensor parallel-based inference and ZeRO-based training sharding mechanisms.

It can also reconfigure the memory system to maximize memory availability in every mode.

This avoids the memory allocation bottleneck and can support large batch sizes, greatly improving performance.

In summary, the Hybrid Engine pushes the boundaries of modern RLHF training, delivering unparalleled scale and system efficiency for RLHF workloads.

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

Effectiveness evaluation

Compared with existing systems such as Colossal-AI or HuggingFace-DDP, DeepSpeed-Chat has over an order of magnitude throughput, enabling training larger actor models at the same latency budget or training similarly sized models at lower cost.

For example, DeepSpeed ​​improves the throughput of RLHF training by more than 10 times on a single GPU. While both CAI-Coati and HF-DDP can run 1.3B models, DeepSpeed ​​can run 6.5B models on the same hardware, which is directly 5 times higher.

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

On multiple GPUs on a single node, DeepSpeed-Chat is 6-19 times faster than CAI-Coati in terms of system throughput, HF- DDP is accelerated by 1.4-10.5 times.

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

The team stated that one of the key reasons why DeepSpeed-Chat can achieve such excellent results is the acceleration provided by the hybrid engine during the generation phase.

Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models

The above is the detailed content of Everyone has a ChatGPT! Microsoft DeepSpeed ​​Chat is shockingly released, one-click RLHF training for hundreds of billions of large models. For more information, please follow other related articles on the PHP Chinese website!

Statement:
This article is reproduced at:51cto.com. If there is any infringement, please contact admin@php.cn delete