Simiao Zhang's starred repositories

TensorRT-LLM

TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.

Language:C++License:Apache-2.0Stargazers:7735Issues:0Issues:0

Megatron-LM

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:9543Issues:0Issues:0

cutlass

CUDA Templates for Linear Algebra Subroutines

Language:C++License:NOASSERTIONStargazers:5015Issues:0Issues:0

cs249r_book

Collaborative book Machine Learning Systems

Language:TeXLicense:NOASSERTIONStargazers:710Issues:0Issues:0

LLMs_interview_notes

该仓库主要记录 大模型(LLMs) 算法工程师相关的面试题

License:Apache-2.0Stargazers:1230Issues:0Issues:0

flash-linear-attention

Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton

Language:PythonLicense:MITStargazers:785Issues:0Issues:0

awesome-llm-powered-agent

Awesome things about LLM-powered agents. Papers / Repos / Blogs / ...

License:MITStargazers:1248Issues:0Issues:0

LLMAgentPapers

Must-read Papers on LLM Agents.

Stargazers:1519Issues:0Issues:0

LLM-Agents-Papers

A repo lists papers related to LLM based agent

Language:PythonStargazers:878Issues:0Issues:0

llama3-from-scratch

llama3 implementation one matrix multiplication at a time

Language:Jupyter NotebookLicense:MITStargazers:11606Issues:0Issues:0

OpenVoice

Instant voice cloning by MyShell.

Language:PythonLicense:MITStargazers:27678Issues:0Issues:0

SWE-agent

SWE-agent takes a GitHub issue and tries to automatically fix it, using GPT-4, or your LM of choice. It solves 12.47% of bugs in the SWE-bench evaluation set and takes just 1 minute to run.

Language:PythonLicense:MITStargazers:12155Issues:0Issues:0

tortoise-tts

A multi-voice TTS system trained with an emphasis on quality

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:12569Issues:0Issues:0

CrossViVit

This repository contains code for the paper "Improving day-ahead Solar Irradiance Time Series Forecasting by Leveraging Spatio-Temporal Context"

Language:PythonLicense:MITStargazers:65Issues:0Issues:0

grok-1

Grok open release

Language:PythonLicense:Apache-2.0Stargazers:49222Issues:0Issues:0

AgentBench

A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)

Language:PythonLicense:Apache-2.0Stargazers:2041Issues:0Issues:0

SWE-bench

[ICLR 2024] SWE-Bench: Can Language Models Resolve Real-world Github Issues?

Language:PythonLicense:MITStargazers:1543Issues:0Issues:0

Open-Sora

Open-Sora: Democratizing Efficient Video Production for All

Language:PythonLicense:Apache-2.0Stargazers:20993Issues:0Issues:0

DiT

Official PyTorch Implementation of "Scalable Diffusion Models with Transformers"

Language:PythonLicense:NOASSERTIONStargazers:5788Issues:0Issues:0
Language:Jupyter NotebookLicense:MITStargazers:948Issues:0Issues:0

torch-discounted-cumsum

Fast Discounted Cumulative Sums in PyTorch

Language:PythonLicense:NOASSERTIONStargazers:94Issues:0Issues:0

flash-attention

Fast and memory-efficient exact attention

Language:PythonLicense:BSD-3-ClauseStargazers:12673Issues:0Issues:0
Language:PythonStargazers:162Issues:0Issues:0

vector-quantize-pytorch

Vector (and Scalar) Quantization, in Pytorch

Language:PythonLicense:MITStargazers:2244Issues:0Issues:0

reformer-pytorch

Reformer, the efficient Transformer, in Pytorch

Language:PythonLicense:MITStargazers:2082Issues:0Issues:0

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Language:PythonLicense:Apache-2.0Stargazers:129987Issues:0Issues:0

performer-pytorch

An implementation of Performer, a linear attention-based transformer, in Pytorch

Language:PythonLicense:MITStargazers:1073Issues:0Issues:0

long-range-arena

Long Range Arena for Benchmarking Efficient Transformers

Language:PythonLicense:Apache-2.0Stargazers:706Issues:0Issues:0

Skyformer

Skyformer: Remodel Self-Attention with Gaussian Kernel and Nystr\"om Method (NeurIPS 2021)

Language:PythonStargazers:53Issues:0Issues:0

google-research

Google Research

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:33533Issues:0Issues:0