microsoft / Megatron-DeepSpeed

Ongoing research training transformer language models at scale, including: BERT & GPT-2

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

load_args_from_checkpoint fails if arguments are missing

Mittagskogel opened this issue · comments

# Args.
if not state_dict:
print_rank_0('Checkpoint not found to provide arguments, using provided arguments.')
return args
if 'args' not in state_dict:
print_rank_0('Checkpoint provided does not have arguments saved, using provided arguments.')
return args

This function should always return two values, see default return in line 529. The function is not behaving as described in the warning messages.

Down the line, this causes a crash when trying to unpack two values in

margs, checkpoint_args = load_args_from_checkpoint(margs)

commented

Hello, have you resolved this issue? @Mittagskogel