xenova / transformers.js

State-of-the-art Machine Learning for the web. Run 🤗 Transformers directly in your browser, with no need for a server!

Home Page:https://huggingface.co/docs/transformers.js

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

GGUF support

Mihaiii opened this issue · comments

Feature request

Right now transformers.js works with ONNX models. It would be useful to also support GGUF files (see llama.cpp)

Motivation

Wider support + ONNX doesn't quantize below 8bit, but GGUF does.

Your contribution

I could help manual testing. Regarding the dev work, I'm unsure.