NVIDIA / TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.

Home Page:https://docs.nvidia.com/deeplearning/transformer-engine/user-guide/index.html

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Could TransformerEngine work with Deepspeed Zero w/ offloading?

leiwen83 opened this issue · comments

Hi,

Since it is common to use with deepspeed zero w/ offloading when training large LLM, does TE currently support in this mode?

Currently deepspeed support is just unittest as refered by TE's readme: microsoft/DeepSpeed#3731

Thx~