transfer_param.py 转换vincuna hf模型成sat模型报错
Lunatic-Solar opened this issue · comments
请安装github最新版sat
git clone https://github.com/THUDM/SwissArmyTransformer
cd SwissArmyTransformer
pip install -e .
这个问题解决了。 但我在finetune vicuna的时候报错了,报错如下:
[图片]
看起来也是sat版本的问题?
cogvlm加载的checkpoint应该是旧版本的llama,如果你想用新版的llama可能需要修改一下cogvlm的代码。
两者的区别是新版sat把mlp内置了,不需要添加mixin。
可能是你转换的不太对吧,转换脚本在这里:https://github.com/THUDM/SwissArmyTransformer/tree/main/examples/llama
eva_args肯定是没有的,因为转换的是语言模型。
cogVLM不是基于vicuna finetune的吗? 我也看到cogVLM的issue提到基于vincuna finetune和基于cogVLM finetune 没有区别。而finutune的代码中是需要eva_args结构里的数据的,那究竟是怎么基于vincuna finetune的呢?有没有可以参考的资料呢
直接用cogvlm的model_config.json加载你转换好的checkpoint文件。也就是说把cogvlm的model_config.json复制到你转换好的权重文件夹里。
如果需要预训练的vit权重,可以把vit部分直接添加到你转换好的checkpoint文件里。
请安装github最新版sat
git clone https://github.com/THUDM/SwissArmyTransformer cd SwissArmyTransformer pip install -e .
请问我下载更新的时候出现了这种错误,您知道如何解决吗?
pip install -e .
Obtaining file:///home/root1/PycharmProjects/Inf-DiT-main/SwissArmyTransformer-main
Preparing metadata (setup.py) ... done
Requirement already satisfied: torch in /home/root1/anaconda3/envs/inf-dit/lib/python3.10/site-packages (from SwissArmyTransformer==0.4.11) (2.3.0)
Collecting deepspeed (from SwissArmyTransformer==0.4.11)
Using cached deepspeed-0.14.2.tar.gz (1.3 MB)
Preparing metadata (setup.py) ... error
error: subprocess-exited-with-error
× python setup.py egg_info did not run successfully.
│ exit code: 1
╰─> [8 lines of output]
Traceback (most recent call last):
File "", line 2, in
File "", line 34, in
File "/tmp/pip-install-woqq3hon/deepspeed_6ad9a8143d084357a3f16a4f3891dace/setup.py", line 100, in
cuda_major_ver, cuda_minor_ver = installed_cuda_version()
File "/tmp/pip-install-woqq3hon/deepspeed_6ad9a8143d084357a3f16a4f3891dace/op_builder/builder.py", line 50, in installed_cuda_version
raise MissingCUDAException("CUDA_HOME does not exist, unable to compile CUDA op(s)")
op_builder.builder.MissingCUDAException: CUDA_HOME does not exist, unable to compile CUDA op(s)
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
error: metadata-generation-failed
× Encountered error while generating package metadata.
╰─> See above for output.
note: This is an issue with the package mentioned above, not pip.
hint: See above for details.
如果不用于训练的话可以这样试试:
pip install . --no-deps
否则就需要安装正确的cuda版本
如果不用于训练的话可以这样试试:
pip install . --no-deps
否则就需要安装正确的cuda版本
好的 谢谢,但是这个对应的cuda版本和python版本如何查看呢
import torch
print(torch.version.cuda)
import torch print(torch.version.cuda)
谢谢你 我尝试一下