Zijie Tian (Zijie-Tian)

Zijie-Tian

Geek Repo

Company:Tsinghua University

Location:Beijing, China

Twitter:@zijie_tian

Github PK Tool:Github PK Tool

Zijie Tian's starred repositories

segment-anything

The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:44992Issues:300Issues:647

jax

Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more

Language:PythonLicense:Apache-2.0Stargazers:28777Issues:326Issues:5247

terraformer

CLI tool to generate terraform files from existing infrastructure (reverse Terraform). Infrastructure to Code

Language:GoLicense:Apache-2.0Stargazers:11989Issues:140Issues:775

ZLUDA

CUDA on AMD GPUs

Language:RustLicense:Apache-2.0Stargazers:8090Issues:114Issues:147

trax

Trax — Deep Learning with Clear Code and Speed

Language:PythonLicense:Apache-2.0Stargazers:7983Issues:146Issues:232

bypy

Python client for Baidu Yun (Personal Cloud Storage) 百度云/百度网盘Python客户端

Language:PythonLicense:MITStargazers:7599Issues:298Issues:567

TensorRT-LLM

TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.

Language:C++License:Apache-2.0Stargazers:7119Issues:82Issues:1445

lit-llama

Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.

Language:PythonLicense:Apache-2.0Stargazers:5860Issues:67Issues:268

script-commands

Script Commands let you tailor Raycast to your needs. Think of them as little productivity boosts throughout your day.

Language:ShellLicense:MITStargazers:5847Issues:46Issues:206
Language:PythonLicense:Apache-2.0Stargazers:4363Issues:47Issues:800

AliceMind

ALIbaba's Collection of Encoder-decoders from MinD (Machine IntelligeNce of Damo) Lab

Language:PythonLicense:Apache-2.0Stargazers:1958Issues:50Issues:78

Efficient-LLMs-Survey

[TMLR 2024] Efficient Large Language Models: A Survey

sparsegpt

Code for the ICML 2023 paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot".

Language:PythonLicense:Apache-2.0Stargazers:653Issues:16Issues:31

PiPPy

Pipeline Parallelism for PyTorch

Language:PythonLicense:BSD-3-ClauseStargazers:651Issues:36Issues:248

llm-inference-benchmark

LLM Inference benchmark

Language:PythonLicense:MITStargazers:265Issues:2Issues:2

local_llama

This repo is to showcase how you can run a model locally and offline, free of OpenAI dependencies.

Language:PythonLicense:Apache-2.0Stargazers:193Issues:6Issues:13

ao

torchao: PyTorch Architecture Optimization (AO). Performant kernels that work with PyTorch.

Language:PythonLicense:BSD-3-ClauseStargazers:192Issues:6Issues:6

reason

A shell for research papers

Language:RustLicense:MITStargazers:189Issues:4Issues:5

gdGPT

Train llm (bloom, llama, baichuan2-7b, chatglm3-6b) with deepspeed pipeline mode. Faster than zero/zero++/fsdp.

Language:PythonLicense:Apache-2.0Stargazers:86Issues:1Issues:8

cuda_scheduling_examiner_mirror

A tool for examining GPU scheduling behavior.

Language:CudaLicense:NOASSERTIONStargazers:62Issues:11Issues:2

Awesome-Resource-Efficient-LLM-Papers

a curated list of high-quality papers on resource-efficient LLMs 🌱

License:CC0-1.0Stargazers:55Issues:5Issues:0

SparseFinetuning

Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry

Language:PythonLicense:Apache-2.0Stargazers:35Issues:5Issues:0

structured_transposable_masks

Code for ICML 2021 submission

UnderstandingNLP

Natural Language Processing Analysis

Language:Jupyter NotebookStargazers:30Issues:4Issues:4

frame

FRAME: Fast Roofline Analytical Modeling and Estimation

Language:Jupyter NotebookLicense:MITStargazers:25Issues:2Issues:0

MLCarbon

End-to-end carbon footprint mod- eling tool

Language:PythonStargazers:24Issues:2Issues:0

Opara

Opara is a lightweight and resource-aware DNN Operator parallel scheduling framework to accelerate the execution of DNN inference on GPUs.

Language:PythonStargazers:16Issues:2Issues:0

model-evaluator

Evaluate Transformers from the Hub 🔥

Language:PythonLicense:Apache-2.0Stargazers:12Issues:8Issues:16

wandbtocsv

CLI tool to export W&B metrics to a csv file.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:10Issues:2Issues:0

lowtime

A time-cost tradeoff problem solver

Language:PythonLicense:Apache-2.0Stargazers:10Issues:3Issues:0