Weigao Sun (weigao266)

weigao266

Geek Repo

Location:Shanghai, China

Github PK Tool:Github PK Tool

Weigao Sun's starred repositories

llama.cpp

LLM inference in C/C++

zed

Code at the speed of thought – Zed is a high-performance, multiplayer code editor from the creators of Atom and Tree-sitter.

Language:RustLicense:NOASSERTIONStargazers:43362Issues:193Issues:7157

google-research

Google Research

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:33578Issues:749Issues:1224

stanford_alpaca

Code and documentation to train Stanford's Alpaca models, and generate the data.

Language:PythonLicense:Apache-2.0Stargazers:29246Issues:340Issues:267

pytorch_geometric

Graph Neural Network Library for PyTorch

Language:PythonLicense:MITStargazers:20738Issues:252Issues:3489

LLaVA

[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.

Language:PythonLicense:Apache-2.0Stargazers:18573Issues:159Issues:1431

ChatPaper

Use ChatGPT to summarize the arXiv papers. 全流程加速科研,利用chatgpt进行论文全文总结+专业翻译+润色+审稿+审稿回复

Language:PythonLicense:NOASSERTIONStargazers:18061Issues:92Issues:216

flash-attention

Fast and memory-efficient exact attention

Language:PythonLicense:BSD-3-ClauseStargazers:12802Issues:114Issues:947

NeMo

A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)

Language:PythonLicense:Apache-2.0Stargazers:11174Issues:201Issues:2182

Awesome-Multimodal-Large-Language-Models

:sparkles::sparkles:Latest Advances on Multimodal Large Language Models

cuda-samples

Samples for CUDA Developers which demonstrates features in CUDA Toolkit

Language:CLicense:NOASSERTIONStargazers:5898Issues:117Issues:229

opencompass

OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.

Language:PythonLicense:Apache-2.0Stargazers:3485Issues:23Issues:443

mamba-minimal

Simple, minimal implementation of the Mamba SSM in one file of PyTorch.

Language:PythonLicense:Apache-2.0Stargazers:2484Issues:23Issues:25

lightllm

LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.

Language:PythonLicense:Apache-2.0Stargazers:2167Issues:23Issues:177

torchmetrics

Torchmetrics - Machine learning metrics for distributed, scalable PyTorch applications.

Language:PythonLicense:Apache-2.0Stargazers:2039Issues:30Issues:848

GenerativeAIExamples

Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:1941Issues:51Issues:34

TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.

Language:PythonLicense:Apache-2.0Stargazers:1707Issues:36Issues:282

Megatron-DeepSpeed

Ongoing research training transformer language models at scale, including: BERT & GPT-2

Language:PythonLicense:NOASSERTIONStargazers:1295Issues:24Issues:143

bigscience

Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.

Language:ShellLicense:NOASSERTIONStargazers:968Issues:38Issues:19

DeepSeek-MoE

DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models

Language:PythonLicense:MITStargazers:935Issues:15Issues:35
Language:PythonLicense:Apache-2.0Stargazers:856Issues:9Issues:0

llama-moe

⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training

Language:PythonLicense:Apache-2.0Stargazers:816Issues:7Issues:18

flash-linear-attention

Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton

Language:PythonLicense:MITStargazers:793Issues:21Issues:31
Language:PythonLicense:Apache-2.0Stargazers:765Issues:12Issues:34

Awesome-Mixture-of-Experts-Papers

A curated reading list of research in Mixture-of-Experts(MoE).

MS-AMP

Microsoft Automatic Mixed Precision Library

Language:PythonLicense:MITStargazers:491Issues:11Issues:61

NeMo-Megatron-Launcher

NeMo Megatron launcher and tools

Language:PythonLicense:Apache-2.0Stargazers:389Issues:19Issues:29

MathPile

Generative AI for Math: MathPile

Language:PythonLicense:Apache-2.0Stargazers:364Issues:7Issues:5

zero-bubble-pipeline-parallelism

Zero Bubble Pipeline Parallelism

Language:PythonLicense:NOASSERTIONStargazers:237Issues:6Issues:23

lightning-attention

Lightning Attention-2: A Free Lunch for Handling Unlimited Sequence Lengths in Large Language Models

Language:PythonLicense:MITStargazers:175Issues:11Issues:12