recmo / cria

Tiny inference-only implementation of LLaMA

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Inference-only implementation of LLaMA in plain NumPy

It uses NumPy, so it can run without a GPU. It also uses memory mapped files to load the weights, so you can run it with little memory.

Inspired by picoGPT.

Besides NumPy, it currently also has a dependency on Google's SentencePiece for tokenization.

About

Tiny inference-only implementation of LLaMA


Languages

Language:Python 100.0%