liuzhuang13 / DenseNet

Densely Connected Convolutional Networks, In CVPR 2017 (Best Paper Award).

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Question about batch_size, epoch setting

guvcolie opened this issue · comments

Hi, I'm trying to train DenseNet-121 on ImageNet dataset, but the result is poor...
Now I wander how to calculate the batch_size in muiti GPUs. You said "It took us 10 days to train 40M densenet for 120 epochs on 4 TITAN X GPUs, with batchsize 128" on issue (https://github.com/liuzhuang13/DenseNet/issues/5), you mean "each GPU use batchsize 128" or "each GPU use 32, sum is 128"?
Thank you!

Thanks and we meant "each GPU use 32, sum is 128".

so... if I set 2 gpus, and batch_size =64, in fact, it is batchsize 128 ?

Actually the "batch_size" in the code means the total batch size. So if you want in total batch size 128 just set batch_size = 128.