Notnaton / microllm

My own implementation to run inference on local LLM models

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Microllm

just the bare basics to run inference on local hardware.

currently working:

  • gguf.py Now it reads the entire gguf file and returns the file locations for the tensor data.

todo:

  • load tensors into model
  • inference

About

My own implementation to run inference on local LLM models

License:GNU Affero General Public License v3.0


Languages

Language:Python 100.0%