With a learning rate of 0.1, loss is 'nan'. I don't kown why?
Hedlen opened this issue · comments
Dylan commented
When using your program,with a learning rate of 0.1, loss is ‘nan’, when I change the learning rate to 0.001, loss changes normally. I don't know why ? According to your documentation, lr 0.1 should be fine.
I look forward to your reply. Thanks.
Bob.Liu commented
hi. how much your label_num (class number)? I had tried 10K and 700K with lr=0.1 and everything is ok.