karpathy / llama2.c

Inference Llama 2 in one file of pure C

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

malloc failed! on stories260 model

vikramdattu opened this issue · comments

I have cloned the tinyllamas repo and trying stories260 model. It fails with the malloc failed issue

./run tinyllamas/stories260K/stories260K.bin -z tinyllamas/stories260K/tok512.bin -t 0.0
malloc failed!

15M model however works out of the box. But not able to run any model from the GH repo!

Trying it on M1 Mac. Any ideas?

When you run tokenizer.py to export the tokenizer, the legacy version will work with run.c