AILab-CVC / SEED

Official implementation of SEED-LLaMA (ICLR 2024).

Home Page:https://ailab-cvc.github.io/seed

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Missing Multimodel Pretraining step

shubhamgarg21 opened this issue · comments

Hi,

For the paper https://arxiv.org/pdf/2310.01218.pdf , the following is mentioned in pretraining section :

For efficiency, we first train SEED-LLaMA using LoRA [32] tuning and together optimize the
parameters of the embedding layer and decoder head layer due to the added visual codes. We then
merge the parameters of LoRA onto the LLM backbone and fine-tune all parameters except for
the embedding layer.

But in the training steps, the part about fine-tuning all parameters except for the embedding layer is missing.