Lukas Kreussel's repositories
llm-rs-python
Unofficial python bindings for the rust llm library. 🐍❤️🦀
Retrieval-Augmented-QA
Query, ask and chat with a document-index via transformer models!
vLLM-haystack-adapter
Simply connect your haystack pipeline to an vLLM-API server
retsim-pytorch
A pytorch port of Google's RETSim model used in UniSim
Google-News-Feed
A simple python library to consume the google news rss feed
Auto-GGML-Conversions
Automated GGML conversions to the Huggingface HUB
candle-vllm
Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.
mistral.rs
Blazingly fast LLM inference.
ReFrontier
Tools for *packing, *crypting and editing various Monster Hunter Frontier Online files.
HunterPie
HunterPie completely rewritten from scratch
continue
⏩ Continue is the leading open-source AI code assistant. You can connect any models and any context to build custom autocomplete and chat experiences inside VS Code and JetBrains
cuda-toolkit
GitHub Action to install CUDA
express
Fast, unopinionated, minimalist web framework for node.
haystack-integrations
🚀 A list of Haystack Integrations, maintained by the community or deepset.
llama-cpp-python
Python bindings for llama.cpp
llama.cpp
Port of Facebook's LLaMA model in C/C++
llm
An ecosystem of Rust libraries for working with large language models
pokerogue
A browser based Pokémon fangame heavily inspired by the roguelite genre.
ratchet
A cross-platform browser ML framework.
rogueserver
Game server backend and API for PokéRogue
SlimeVR-Tracker-ESP-BMI270
SlimeVR tracker firmware for ESP32/ESP8266 and different IMUs
st-chat
Streamlit Component, for a Chatbot UI
ST4-Python-Parser
Parse Schema ST4 XML Files into simple, flat python objects.
tabby
Self-hosted AI coding assistant
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
TTS
🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
wonnx
A WebGPU-accelerated ONNX inference run-time written 100% in Rust, ready for native and the web