CUDA out of memory when validation
ArtemisZGL opened this issue · comments
First i want to thanks for your work. And I run you code with dataset llibrispeech train 100 as training data, dev clean as validation data using GeForce RTX 2070. After several time OOM error, i set the barch size to 4 and finally can train normally. But after one epoch, i also met the OOM error in validation. So i want to konw if i set the batch size smaller can avoid this problem ? Because i notice that in the librispeech dataset process script, the training data have been pruned to min/max duration but the validation and test data didn't.
And I also to konw is there a result for librispeech using this code ? I only saw the aishell result in README.
Thanks.
I added with torch.no_grad():
before the validation loop:
end2end-asr-pytorch/trainer/asr/trainer.py
Line 127 in a22efdd
with torch.no_grad():
, Line 31 in a22efdd
@paanguin thanks! i will try this.
thanks @paanguin I am closing the issue