stas00 / porting

Helper scripts and notes that were used while porting various nlp models

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Disable extreme deepspeed logging

iliaschalkidis opened this issue · comments

Hi @stas00,

Thanks for the deepspeed notebook example. Is there any way to disable deespeed logging?

I mean all these repetive lines, that go in between the usual logging of huggingface training script:

rank=0 time (ms) | optimizer_gradients: 51.24 | optimizer_step: 937.01 | optimizer_allgather: 64.85
[2021-03-10 10:46:04,811] [INFO] [logging.py:60:log_dist] [Rank 0] step=710, skipped=19, lr=[2.359596026758678e-05], mom=[[0.9, 0.999]]
[2021-03-10 10:46:04,811] [INFO] [timer.py:163:stop] 0/710, SamplesPerSec=174.80488942519466

Especially the optimizer info, which is printed constantly 😄

I totally agree, @iliaschalkidis!

In general, pretty much any DeepSpeed-specific questions should go to https://github.com/microsoft/DeepSpeed - please feel free to tag me if it's related to transformers though, since most likely I need to know it too.

In this particular situation please watch for the resolution of this PR microsoft/DeepSpeed#849

Currently, those timing prints are controlled by:

    "wall_clock_breakdown": false

in the DS config file - and it should be false by default - except it's not working right now and this PR I linked to will resolve it.

Thanks @stas00,

I already added this in the config:

"steps_per_print": 5000,
"wall_clock_breakdown": false,

but it seems it only suppress some of the logging, not everything...

I'll keep up for updates!

It should be good now, just for wait for the PR to be merged and use master or the upcoming new release.