Zhi-guo Huang (hzg0601)

hzg0601

Geek Repo

Company:pinming

Location:Hangzhou,China

Home Page:https://hzg0601.github.io/

Twitter:@huang_zhiguo

Github PK Tool:Github PK Tool

Zhi-guo Huang's repositories

LLM-Notes

大模型技术栈一览

langchain-ChatGLM-annotation

对langchain-ChatGLM项目各模块进行注释,增加了一些新的特性,修复了一些bug

Language:PythonLicense:Apache-2.0Stargazers:23Issues:0Issues:0

cn-llm-codes

中文LLM的代码合集

Language:PythonStargazers:2Issues:1Issues:0

speedai

大规模AI加速方法笔记

Language:Jupyter NotebookStargazers:1Issues:0Issues:0

mii-dev

dev for deepspeed-mii

Language:PythonLicense:Apache-2.0Stargazers:1Issues:1Issues:0
Language:PythonLicense:MITStargazers:1Issues:1Issues:0

chat-gpt-langchain-fork

fork from https://huggingface.co/spaces/JavaFXpert/Chat-GPT-LangChain

Language:PythonStargazers:0Issues:1Issues:0

debuged-Evolve-GCN

对evolve-gcn的源代码进行了debug

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

ds-chat-bloom

ds-chat 针对bloom进行了debug

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:HTMLLicense:MITStargazers:0Issues:1Issues:0

lit-llama-cn-annotated

Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

peft-cn-annotated

🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

qlora-zero-cn

qlora+zero算法,加速模型训练,降低显存要求; qlora各模块的中文注释

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:0Issues:0Issues:0

ZeQLoRA

ZeQLoRA: Efficient Finetuning of Quantized LLMs with ZeRO and LoRA

Language:Jupyter NotebookLicense:MITStargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:1Issues:0

clash-for-linux-backup

clash for linux备份仓库

Language:ShellStargazers:0Issues:0Issues:0

cs-224w-cn

cs224w课程的中文笔记

Language:Jupyter NotebookStargazers:0Issues:0Issues:0

DeepKE-fork

An Open Toolkit for Knowledge Graph Extraction and Construction published at EMNLP2022 System Demonstrations.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

Fast-Chatchat

FastChat中文注释见cn_annotation分支,新特性见readme

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

fastllm-fork

纯c++的全平台llm加速库,支持python调用,chatglm-6B级模型单卡可达10000+token / s,支持glm, llama, moss基座,手机端流畅运行

Language:C++Stargazers:0Issues:0Issues:0
Stargazers:0Issues:0Issues:0

GPTCache-dev

Semantic cache for LLMs. Fully integrated with LangChain and llama_index.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

graphrag-fork

A modular graph-based Retrieval-Augmented Generation (RAG) system

License:MITStargazers:0Issues:0Issues:0

inference-dev

Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

LightChat

一个用于提供LLM服务的轻量级工具

Language:PythonStargazers:0Issues:1Issues:0

Megatron-LM-fork

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

TensorRT-LLM-dev

TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.

License:Apache-2.0Stargazers:0Issues:0Issues:0

weaviate-abc

weaviate入门

Language:PythonStargazers:0Issues:0Issues:0