microsoft / Megatron-DeepSpeed

Ongoing research training transformer language models at scale, including: BERT & GPT-2

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

[QUESTION] Does the dev team have a plan to merge Mega-LM 0.4?

nrailg opened this issue · comments

Last week Nvidia released Mega-LM 0.4. It greatly enhanced code / module organization, integrates transformer_engine, and etc.

Does the dev team have a plan to merge Mega-LM 0.4? especially the transformer_engine?