ggerganov / llama.cpp

LLM inference in C/C++

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

An error occurred while converting Sakura-14B-Qwen2beta-v0.10pre0 to gguf

lingyezhixing opened this issue · comments

Traceback (most recent call last):
File "D:\LLM\llama.cpp\convert.py", line 1584, in
main()
File "D:\LLM\llama.cpp\convert.py", line 1578, in main
OutputFile.write_all(outfile, ftype, params, model, vocab, special_vocab,
File "D:\LLM\llama.cpp\convert.py", line 1240, in write_all
of.write_tensor_data(ftype, model, concurrency)
File "D:\LLM\llama.cpp\convert.py", line 1178, in write_tensor_data
self.gguf.write_tensor_data(ndarray)
File "D:\LLM\llama.cpp\gguf-py\gguf\gguf_writer.py", line 272, in write_tensor_data
tensor.tofile(self.fout)
OSError: 70123520 requested and 0 written

This model is a translation model based on qwen2
Model address:https://hf-mirror.com/SakuraLLM/Sakura-14B-Qwen2beta-v0.10pre0