MIT HAN Lab (mit-han-lab)

MIT HAN Lab

mit-han-lab

Geek Repo

Efficient AI Computing. PI: Song Han

Location:MIT

Home Page:https://hanlab.mit.edu

Twitter:@songhan_mit

Github PK Tool:Github PK Tool

MIT HAN Lab's repositories

streaming-llm

[ICLR 2024] Efficient Streaming Language Models with Attention Sinks

Language:PythonLicense:MITStargazers:6350Issues:60Issues:78

bevfusion

[ICRA'23] BEVFusion: Multi-Task Multi-Sensor Fusion with Unified Bird's-Eye View Representation

Language:PythonLicense:Apache-2.0Stargazers:2127Issues:42Issues:603

llm-awq

[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration

Language:PythonLicense:MITStargazers:2073Issues:23Issues:157

temporal-shift-module

[ICCV 2019] TSM: Temporal Shift Module for Efficient Video Understanding

Language:PythonLicense:MITStargazers:2034Issues:42Issues:219

once-for-all

[ICLR 2020] Once for All: Train One Network and Specialize it for Efficient Deployment

Language:PythonLicense:MITStargazers:1848Issues:53Issues:75

efficientvit

EfficientViT is a new family of vision models for efficient high-resolution vision.

Language:PythonLicense:Apache-2.0Stargazers:1588Issues:33Issues:114

data-efficient-gans

[NeurIPS 2020] Differentiable Augmentation for Data-Efficient GAN Training

Language:PythonLicense:BSD-2-ClauseStargazers:1268Issues:19Issues:97

torchquantum

A PyTorch-based framework for Quantum Classical Simulation, Quantum Machine Learning, Quantum Neural Networks, Parameterized Quantum Circuits with support for easy deployments on real quantum computers.

Language:Jupyter NotebookLicense:MITStargazers:1236Issues:25Issues:115

torchsparse

[MICRO'23, MLSys'22] TorchSparse: Efficient Training and Inference Framework for Sparse Convolution on GPUs.

Language:CudaLicense:MITStargazers:1149Issues:18Issues:239

gan-compression

[CVPR 2020] GAN Compression: Efficient Architectures for Interactive Conditional GANs

Language:PythonLicense:NOASSERTIONStargazers:1098Issues:29Issues:101

smoothquant

[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models

Language:PythonLicense:MITStargazers:1098Issues:19Issues:81

anycost-gan

[CVPR 2021] Anycost GANs for Interactive Image Synthesis and Editing

Language:PythonLicense:MITStargazers:775Issues:23Issues:30

tinyengine

[NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning; [NeurIPS 2022] MCUNetV3: On-Device Training Under 256KB Memory

Language:PythonLicense:MITStargazers:721Issues:37Issues:26

fastcomposer

FastComposer: Tuning-Free Multi-Subject Image Generation with Localized Attention

Language:PythonLicense:MITStargazers:621Issues:21Issues:32

TinyChatEngine

TinyChatEngine: On-Device LLM Inference Library

Language:C++License:MITStargazers:621Issues:12Issues:37

distrifuser

[CVPR 2024 Highlight] DistriFusion: Distributed Parallel Inference for High-Resolution Diffusion Models

Language:PythonLicense:MITStargazers:487Issues:8Issues:16

mcunet

[NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning

Language:PythonLicense:MITStargazers:422Issues:23Issues:27

amc

[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices

Language:PythonLicense:MITStargazers:421Issues:17Issues:25

tiny-training

On-Device Training Under 256KB Memory [NeurIPS'22]

Language:PythonLicense:MITStargazers:414Issues:17Issues:8

offsite-tuning

Offsite-Tuning: Transfer Learning without Full Model

Language:PythonLicense:MITStargazers:363Issues:8Issues:10

qserve

QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving

Language:PythonLicense:Apache-2.0Stargazers:325Issues:8Issues:16

litepose

[CVPR'22] Lite Pose: Efficient Architecture Design for 2D Human Pose Estimation

Language:PythonLicense:MITStargazers:299Issues:23Issues:37

flatformer

[CVPR'23] FlatFormer: Flattened Window Attention for Efficient Point Cloud Transformer

Language:PythonLicense:Apache-2.0Stargazers:75Issues:1Issues:9

Quest

[ICML 2024] Quest: Query-Aware Sparsity for Efficient Long-Context LLM Inference

patch_conv

Patch convolution to avoid large GPU memory usage of Conv2D

Language:PythonLicense:MITStargazers:67Issues:8Issues:1

sparsevit

[CVPR'23] SparseViT: Revisiting Activation Sparsity for Efficient High-Resolution Vision Transformer

Language:PythonLicense:Apache-2.0Stargazers:56Issues:4Issues:2

spatten-llm

[HPCA'21] SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning

Language:ScalaLicense:MITStargazers:55Issues:8Issues:1