Jan's repositories
awesome-local-ai
An awesome repository of local AI tools
nitro-tensorrt-llm
Nitro is an C++ inference server on top of TensorRT-LLM. OpenAI-compatible API. Run blazing fast inference on Nvidia GPUs. Used in Jan
cortex.python-runtime
C++ code that run Python embedding
thinking-machines
Thinking Machines
open-foundry
R&D experiments
Real-ESRGAN
Real-ESRGAN aims at developing Practical Algorithms for General Image/Video Restoration.
infinity
The AI-native database built for LLM applications, providing incredibly fast vector and full-text search
llama.cpp-avx-vnni
Port of Facebook's LLaMA model in C/C++
openai_trtllm
OpenAI compatible API for TensorRT LLM triton backend
pymaker
Make the py
tensorrtllm_backend
The Triton TensorRT-LLM Backend
trt-llm-as-openai-windows
This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows instead of cloud.