Ruonan Wang (rnwang04)

rnwang04

Geek Repo

Company:Intel

Location:Shanghai, China

Github PK Tool:Github PK Tool

Ruonan Wang's starred repositories

ollama

Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.

annotated_deep_learning_paper_implementations

🧑‍🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠

Language:PythonLicense:MITStargazers:54386Issues:448Issues:132

text-generation-webui

A Gradio web UI for Large Language Models.

Language:PythonLicense:AGPL-3.0Stargazers:39811Issues:327Issues:3619

LLaMA-Factory

Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)

Language:PythonLicense:Apache-2.0Stargazers:31740Issues:201Issues:4905

faiss

A library for efficient similarity search and clustering of dense vectors.

milvus

A cloud-native vector database, storage for next generation AI applications

Language:GoLicense:Apache-2.0Stargazers:29681Issues:277Issues:11816

LLMs-from-scratch

Implementing a ChatGPT-like LLM in PyTorch from scratch, step by step

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:28003Issues:301Issues:89

vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

Language:PythonLicense:Apache-2.0Stargazers:27640Issues:226Issues:4646

jan

Jan is an open source alternative to ChatGPT that runs 100% offline on your computer. Multiple engine support (llama.cpp, TensorRT-LLM)

Language:TypeScriptLicense:AGPL-3.0Stargazers:22443Issues:126Issues:1766

Awesome-pytorch-list

A comprehensive list of pytorch related content on github,such as different models,implementations,helper libraries,tutorials etc.

onnxruntime

ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator

axolotl

Go ahead and axolotl questions

Language:PythonLicense:Apache-2.0Stargazers:7649Issues:49Issues:649

Qwen2

Qwen2 is the large language model series developed by Qwen team, Alibaba Cloud.

chainlit

Build Conversational AI in minutes ⚡️

Language:PythonLicense:Apache-2.0Stargazers:6872Issues:53Issues:769

gpt-fast

Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.

Language:PythonLicense:BSD-3-ClauseStargazers:5559Issues:63Issues:98

koboldcpp

Run GGUF models easily with a KoboldAI UI. One File. Zero Install.

Language:C++License:AGPL-3.0Stargazers:4950Issues:69Issues:750

llm-numbers

Numbers every LLM developer should know

cortex

Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM, ONNX). Powers 👋 Jan

Language:C++License:Apache-2.0Stargazers:1885Issues:14Issues:411

LLaVA-Med

Large Language-and-Vision Assistant for Biomedicine, built towards multimodal GPT-4 level capabilities.

Language:PythonLicense:NOASSERTIONStargazers:1475Issues:28Issues:85

onnxruntime-inference-examples

Examples for using ONNX Runtime for machine learning inferencing.

Language:C++License:MITStargazers:1155Issues:38Issues:156
Language:PythonLicense:GPL-3.0Stargazers:486Issues:12Issues:62

intel-npu-acceleration-library

Intel® NPU Acceleration Library

Language:PythonLicense:Apache-2.0Stargazers:447Issues:29Issues:71

chatllm.cpp

Pure C++ implementation of several models for real-time chatting on your computer (CPU)

Language:C++License:MITStargazers:356Issues:17Issues:35
Language:LLVMLicense:NOASSERTIONStargazers:221Issues:14Issues:129

level-zero

oneAPI Level Zero Specification Headers and Loader

Language:C++License:MITStargazers:209Issues:33Issues:136

fp6_llm

An efficient GPU support for LLM inference with x-bit quantization (e.g. FP6,FP5).

Language:CudaLicense:Apache-2.0Stargazers:179Issues:5Issues:9

linux-npu-driver

Intel® NPU (Neural Processing Unit) Driver

Language:C++License:MITStargazers:151Issues:12Issues:17

Data-Paralle-Cpp

个人翻译《Data Parallel C++》

Language:TeXLicense:Apache-2.0Stargazers:68Issues:3Issues:0

Langchain-Chatchat

Knowledge Base QA using RAG pipeline on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) with IPEX-LLM

Language:PythonLicense:Apache-2.0Stargazers:14Issues:1Issues:0