Different batchsize settings lead to different results
guggugg opened this issue · comments
Hello, thank you for sharing. I trained this code on the LibriMix dataset. When the batchsize=1, the loss can drop normally, but when the batchsize=4, the loss is always positive. What could be the problem, can you give me some help?
Thanks for reaching out. What does it mean drops normally? With batch size of 1 the results should be worse since with batch size of 1 there is no augmentation but I do not understand the always positive part.
If you use other trainers from other repos I can't possibly know what goes wrong.