Alex (CanyonWind)

CanyonWind

Geek Repo

Location:Bay area

Github PK Tool:Github PK Tool

Alex's starred repositories

Language:PythonLicense:Apache-2.0Stargazers:404Issues:0Issues:0

MoE-LLaVA

Mixture-of-Experts for Large Vision-Language Models

Language:PythonLicense:Apache-2.0Stargazers:1945Issues:0Issues:0

web-llm

High-performance In-browser LLM Inference Engine

Language:TypeScriptLicense:Apache-2.0Stargazers:13343Issues:0Issues:0

ml-engineering

Machine Learning Engineering Open Book

Language:PythonLicense:CC-BY-SA-4.0Stargazers:11425Issues:0Issues:0

llama-moe

⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training (EMNLP 2024)

Language:PythonLicense:Apache-2.0Stargazers:872Issues:0Issues:0

mistral-inference

Official inference library for Mistral models

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:9639Issues:0Issues:0

moe_attention

Official repository for the paper "SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention"

Language:PythonLicense:MITStargazers:89Issues:0Issues:0

oneflow

OneFlow is a deep learning framework designed to be user-friendly, scalable and efficient.

Language:C++License:Apache-2.0Stargazers:5884Issues:0Issues:0

dilated-attention-pytorch

(Unofficial) Implementation of dilated attention from "LongNet: Scaling Transformers to 1,000,000,000 Tokens" (https://arxiv.org/abs/2307.02486)

Language:PythonLicense:MITStargazers:50Issues:0Issues:0

dilated-self-attention

Implementation of the dilated self attention as described in "LongNet: Scaling Transformers to 1,000,000,000 Tokens"

Language:PythonLicense:MITStargazers:13Issues:0Issues:0

TorchIntegral

Integral Neural Networks in PyTorch

Language:PythonLicense:Apache-2.0Stargazers:121Issues:0Issues:0

Segment-and-Track-Anything

An open-source project dedicated to tracking and segmenting any objects in videos, either automatically or interactively. The primary algorithms utilized include the Segment Anything Model (SAM) for key-frame segmentation and Associating Objects with Transformers (AOT) for efficient tracking and propagation purposes.

Language:Jupyter NotebookLicense:AGPL-3.0Stargazers:2819Issues:0Issues:0

AutoGPT

AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.

Language:PythonLicense:NOASSERTIONStargazers:167607Issues:0Issues:0

tomesd

Speed up Stable Diffusion with this one simple trick!

Language:PythonLicense:MITStargazers:1274Issues:0Issues:0

web-stable-diffusion

Bringing stable diffusion models to web browsers. Everything runs inside the browser with no server support.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:3567Issues:0Issues:0

tuning_playbook

A playbook for systematically maximizing the performance of deep learning models.

License:NOASSERTIONStargazers:26894Issues:0Issues:0

nni

An open source AutoML toolkit for automate machine learning lifecycle, including feature engineering, neural architecture search, model compression and hyper-parameter tuning.

Language:PythonLicense:MITStargazers:14021Issues:0Issues:0

FasterTransformer

Transformer related optimization, including BERT, GPT

Language:C++License:Apache-2.0Stargazers:5828Issues:0Issues:0

MegCC

MegCC是一个运行时超轻量,高效,移植简单的深度学习模型编译器

Language:C++License:Apache-2.0Stargazers:473Issues:0Issues:0

dmls-book

Summaries and resources for Designing Machine Learning Systems book (Chip Huyen, O'Reilly 2022)

Stargazers:2255Issues:0Issues:0

TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.

Language:PythonLicense:Apache-2.0Stargazers:1883Issues:0Issues:0

GenNAS

Generic Neural Architecture Search via Regression (NeurIPS'21 Spotlight)

Language:PythonStargazers:36Issues:0Issues:0

xformers

Hackable and optimized Transformers building blocks, supporting a composable construction.

Language:PythonLicense:NOASSERTIONStargazers:8489Issues:0Issues:0

FLASHQuad_pytorch

FLASHQuad_pytorch

Language:PythonLicense:MITStargazers:66Issues:0Issues:0

diffusion_distiller

🚀 PyTorch Implementation of "Progressive Distillation for Fast Sampling of Diffusion Models(v-diffusion)"

Language:PythonLicense:MITStargazers:216Issues:0Issues:0

FLASH-pytorch

Implementation of the Transformer variant proposed in "Transformer Quality in Linear Time"

Language:PythonLicense:MITStargazers:345Issues:0Issues:0

TensorRT

PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT

Language:PythonLicense:BSD-3-ClauseStargazers:2565Issues:0Issues:0

guided-inpainting

Towards Unified Keyframe Propagation Models

Language:PythonLicense:MITStargazers:233Issues:0Issues:0

einops

Flexible and powerful tensor operations for readable and reliable code (for pytorch, jax, TF and others)

Language:PythonLicense:MITStargazers:8444Issues:0Issues:0

flash-attention

Fast and memory-efficient exact attention

Language:PythonLicense:BSD-3-ClauseStargazers:13804Issues:0Issues:0