mistralai / mistral-inference

Official inference library for Mistral models

Home Page:https://mistral.ai/

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Missing the params.json

littlewwwhite opened this issue · comments

FileNotFoundError: [Errno 2] No such file or directory: 'Mistral-7B-Instruct-v0.2/params.json'

and I replace the params.json to config.json but still lack of several params such like kv_cache dim etc.

Same here. Also, the model on huggingface has different names for w1/w2/w3 weights (up_proj, down_proj, gate_proj), and it's not documented anywhere. I guess the model from HF works with the Transformers library, but both this and MLX impl need fixing to use it.

Just add the different ones to your params.json file for w1/w2/w3 etc…