kmcgrie's repositories
lc0
SYCL work
intel-extension-for-transformers
⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
triton
Development repository for the Triton language and compiler
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
llvm
Intel staging area for llvm.org contribution. Home for Intel LLVM-based projects.
llama.cpp
Port of Facebook's LLaMA model in C/C++