Xwin-LM / Xwin-LM

Xwin-LM: Powerful, Stable, and Reproducible LLM Alignment

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Can this model inference/quantize using llama-cpp?

Adawann opened this issue · comments

error loading model: create_tensor: tensor 'output_norm.weight' not found
llama_load_model_from_file: failed to load model
llama_init_from_gpt_params: error: failed to load model './models/Xwin-LM-70B-V0.1/ggml-model-f32.gguf'
main: error: unable to load model

cannot run infer :(

this seems to be your specific implementation problem. I tried working with Xwin LMs and they work great. I just used LM studio

@asfandsaleem how did you get it to work on lm studio? I tried using the Vicuna preset without success