sail-sg / lorahub

[COLM 2024] LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Composition

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

train_model model load repaired

JornyWan opened this issue · comments

In the code train_model.py:
model = AutoModelForSeq2SeqLM.from_pretrained(
model_args.model_name_or_path,
from_tf=bool(".ckpt" in model_args.model_name_or_path),
config=config,
cache_dir=model_args.cache_dir,
revision=model_args.model_revision,
use_auth_token=True if model_args.use_auth_token else None,
)
if anyone could not use this to initial a flan-t5 model from AutoModelForSeq2SeqLM, then you need to add the following in params:
unk_token="",
bos_token="",
eos_token="
"
Thanks!

@JornyWan Thanks for your feedback and pull request! May I know your transformers version? I have never encountered the problem, not sure if it is the case for the latest transformers library.

@SivilTaram thanks for quick response, my transformer version is 4.31.0

actually it would be like:
image

@SivilTaram you can just put it to another branch for different version of transformer use if it is the problem with version