Hanry's starred repositories
mailcow-dockerized
mailcow: dockerized - 🐮 + 🐋 = 💕
LLMSys-PaperList
Large Language Model (LLM) Systems Paper List
stable-fast
Best inference performance optimization framework for HuggingFace Diffusers on NVIDIA GPUs.
TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
open-interpreter
A natural language interface for computers
ipex-llm
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, Axolotl, etc.
running_page
Make your own running home page