Bing Han's repositories
CMakeTutorial
CMake中文实战教程
optimum-ascend
Optimized inference with Ascend and Hugging Face
fast-chatglm
Faster ChatGLM-6B with CTranslate2
Ascend-text-generation-inference
huggingface/text-generation-inference 适配昇腾NPU
pytorch-npu
Ascend PyTorch adapter (torch_npu). Mirror of https://gitee.com/ascend/pytorch
elasticsearch-jieba-plugin
jieba analysis plugin for elasticsearch 7.0.0, 6.4.0, 6.0.0, 5.4.0,5.3.0, 5.2.2, 5.2.1, 5.2, 5.1.2, 5.1.1
ChatGLM-Efficient-Tuning
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
CTranslate2
Fast inference engine for Transformer models
faster-whisper
Faster Whisper transcription with CTranslate2
fastllm
纯c++的全平台llm加速库,支持python调用,chatglm-6B级模型单卡可达10000+token / s,支持glm, llama, moss基座,手机端流畅运行
langchain
🦜🔗 Build context-aware reasoning applications
langchain-ChatGLM
langchain-ChatGLM, local knowledge based ChatGLM with langchain | 基于本地知识的 ChatGLM 问答
lightllm
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.
nanoRWKV
The nanoGPT-style implementation of RWKV Language Model - an RNN with GPT-level LLM performance.
NvChad
Blazing fast Neovim config providing solid defaults and a beautiful UI, enhancing your neovim experience.
nvchad-starter
Starter config for NvChad
optimum
🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools
ragflow
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
rwkv.c
Inference Llama 2 in one file of pure C
speaker-verification
speaker verification using pyannote
ssr-command-client
:airplane:The commend client of ssr based Python3
SwissArmyTransformer
SwissArmyTransformer is a flexible and powerful library to develop your own Transformer variants.
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs