Mikasa (jsw-zorro)

jsw-zorro

Geek Repo

Company:University of Michigan

Home Page:shuoweijin.com

Twitter:@shuoweijin

Github PK Tool:Github PK Tool

Mikasa's starred repositories

CitationMap

A simple pip-installable Python tool to generate your own HTML citation world map from your Google Scholar ID.

Language:PythonLicense:NOASSERTIONStargazers:223Issues:0Issues:0

sglang

SGLang is yet another fast serving framework for large language models and vision language models.

Language:PythonLicense:Apache-2.0Stargazers:3585Issues:0Issues:0

ServerlessLLM

Cost-efficient and fast multi-LLM serving.

Language:PythonStargazers:118Issues:0Issues:0

how-to-learn-deep-learning-framework

how to learn PyTorch and OneFlow

License:Apache-2.0Stargazers:297Issues:0Issues:0

pytorch-cppcuda-tutorial

tutorial for writing custom pytorch cpp+cuda kernel, applied on volume rendering (NeRF)

Language:CudaStargazers:359Issues:0Issues:0

lectures

Material for cuda-mode lectures

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:2001Issues:0Issues:0

vidur

A large-scale simulation framework for LLM inference

Language:PythonLicense:MITStargazers:164Issues:0Issues:0

sarathi-serve

A low-latency & high-throughput serving engine for LLMs

Language:PythonLicense:Apache-2.0Stargazers:118Issues:0Issues:0

calculate-flops.pytorch

The calflops is designed to calculate FLOPs、MACs and Parameters in all various neural networks, such as Linear、 CNN、 RNN、 GCN、Transformer(Bert、LlaMA etc Large Language Model)

Language:PythonLicense:MITStargazers:427Issues:0Issues:0

DistServe

Disaggregated serving system for Large Language Models (LLMs).

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:211Issues:0Issues:0

torchtitan

A native PyTorch Library for large model training

Language:PythonLicense:BSD-3-ClauseStargazers:1377Issues:0Issues:0

long-context-attention

Sequence Parallel Attention for Long Context LLM Model Training and Inference

Language:PythonStargazers:241Issues:0Issues:0

qserve

QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving

Language:PythonLicense:Apache-2.0Stargazers:360Issues:0Issues:0

llama3-from-scratch

llama3 implementation one matrix multiplication at a time

Language:Jupyter NotebookLicense:MITStargazers:11601Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:35Issues:0Issues:0

guidance

A guidance language for controlling large language models.

Language:Jupyter NotebookLicense:MITStargazers:18339Issues:0Issues:0

S-LoRA

S-LoRA: Serving Thousands of Concurrent LoRA Adapters

Language:PythonLicense:Apache-2.0Stargazers:1654Issues:0Issues:0

prometheus-eval

Evaluate your LLM's response with Prometheus and GPT4 💯

Language:PythonLicense:Apache-2.0Stargazers:706Issues:0Issues:0

attorch

A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.

Language:PythonLicense:MITStargazers:425Issues:0Issues:0

llama3

The official Meta Llama 3 GitHub site

Language:PythonLicense:NOASSERTIONStargazers:24966Issues:0Issues:0

torchtune

A Native-PyTorch Library for LLM Fine-tuning

Language:PythonLicense:BSD-3-ClauseStargazers:3693Issues:0Issues:0
Language:PythonLicense:MITStargazers:1429Issues:0Issues:0

VAR

[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly yet state-of-the-art* codebase for autoregressive image generation!

Language:PythonLicense:MITStargazers:3894Issues:0Issues:0

llm.c

LLM training in simple, raw C/CUDA

Language:CudaLicense:MITStargazers:22393Issues:0Issues:0

Primo

Primo: Practical Learning-Augmented Systems with Interpretable Models

Language:JavaScriptLicense:Apache-2.0Stargazers:17Issues:0Issues:0

outlines

Structured Text Generation

Language:PythonLicense:Apache-2.0Stargazers:7414Issues:0Issues:0

SWE-agent

SWE-agent takes a GitHub issue and tries to automatically fix it, using GPT-4, or your LM of choice. It solves 12.47% of bugs in the SWE-bench evaluation set and takes just 1 minute to run.

Language:PythonLicense:MITStargazers:12155Issues:0Issues:0

scattermoe

Triton-based implementation of Sparse Mixture of Experts.

Language:PythonLicense:Apache-2.0Stargazers:152Issues:0Issues:0

LLM-Blender

[ACL2023] We introduce LLM-Blender, an innovative ensembling framework to attain consistently superior performance by leveraging the diverse strengths of multiple open-source LLMs. LLM-Blender cut the weaknesses through ranking and integrate the strengths through fusing generation to enhance the capability of LLMs.

Language:PythonLicense:Apache-2.0Stargazers:836Issues:0Issues:0

Awesome-Efficient-LLM

A curated list for Efficient Large Language Models

Language:PythonStargazers:1004Issues:0Issues:0