yaoyao-liu / meta-transfer-learning

TensorFlow and PyTorch implementation of "Meta-Transfer Learning for Few-Shot Learning" (CVPR2019)

Home Page:https://lyy.mpi-inf.mpg.de/mtl/

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

the choice of optimizer

Sword-keeper opened this issue · comments

Hi I just found that in the pretrain phase, you choose the SGD optimizer. However, in the meta-train phase, you choose the Adam optimizer. I wonder that why you choose different optimizer in the different phase?

We choose the optimizer by empirical results. You may change the optimizer and re-run the experiments to see the difference.

hi In the torch code,I found that you set the train_aug=false in the meta-training phase.However, in the pretrain phase, you set the train_aug = true. So the train_aug is designed for the pretrain phase? I set train_aug=ture in the meta-training phase, and runned several epochs. The result lower than aug=False.

We apply data augmentation during pre-training to solve the overfitting problem. You may also apply data augmentation during meta-training as well. Please note that you cannot apply data augmentation on the episode test (the test set for each small task).