piano_123's starred repositories
RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
gaussian-splatting
Original reference implementation of "3D Gaussian Splatting for Real-Time Radiance Field Rendering"
stable-diffusion-webui
Stable Diffusion web UI
x-transformers
A simple but complete full-attention transformer with a set of promising experimental features from various papers
Video-LLaMA
[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
tensorrtllm_backend
The Triton TensorRT-LLM Backend
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
CUDALibrarySamples
CUDA Library Samples
GPTQ-for-LLaMa
4 bits quantization of LLaMA using GPTQ
ChatGLM-Efficient-Tuning
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
LangChain-Chinese-Getting-Started-Guide
LangChain 的中文入门教程
bitsandbytes
Accessible large language models via k-bit quantization for PyTorch.