hppanev's starred repositories
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
FlagEmbedding
Retrieval and Retrieval-augmented LLMs
RWKV-Runner
A RWKV management and startup tool, full automation, only 8MB. And provides an interface compatible with the OpenAI API. RWKV is a large language model that is fully open source and available for commercial use.
human-eval
Code for the paper "Evaluating Large Language Models Trained on Code"
GenerativeAIExamples
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
Stable-Diffusion-WebUI-TensorRT
TensorRT Extension for Stable Diffusion Web UI
ComfyUI-InstantID
Unofficial implementation of InstantID for ComfyUI
GPU-Benchmarks-on-LLM-Inference
Multiple NVIDIA GPUs or Apple Silicon for Large Language Model Inference?
llm-autoeval
Automatically evaluate your LLMs in Google Colab
dark-background-light-text-extension
Firefox addon that turns every page colors into "light text on dark background"
Lora-Training-in-Comfy
This custom node lets you train LoRA directly in ComfyUI!
ComfyUI-InstantID
Thin Custom Node wrapper for InstantID in ComfyUI.