TRL documentation

DeepSpeed Integration

You are viewing main version, which requires installation from source. If you'd like regular pip install, checkout the latest stable version (v0.15.2).
Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

DeepSpeed Integration

Section under construction. Feel free to contribute!

TRL supports training with DeepSpeed, a library that implements advanced training optimization techniques. These include optimizer state partitioning, offloading, gradient partitioning, and more.

DeepSpeed integrates the Zero Redundancy Optimizer (ZeRO), which allows to scale the model size proportional to the number of devices with sustained high efficiency.

ZeRO Stages

Installation

To use DeepSpeed with TRL, install it using the following command:

pip install deepspeed

Running Training Scripts with DeepSpeed

No modifications to your training script are required. Simply run it with the DeepSpeed configuration file:

accelerate launch --config_file <ACCELERATE_WITH_DEEPSPEED_CONFIG_FILE.yaml> train.py

We provide ready-to-use DeepSpeed configuration files in the examples/accelerate_configs directory. For example, to run training with ZeRO Stage 2, use the following command:

accelerate launch --config_file examples/accelerate_configs/deepspeed_zero2.yaml train.py

Additional Resources

Consult the 🤗 Accelerate documentation for more information about the DeepSpeed plugin.

< > Update on GitHub