alphaRGB's starred repositories
activitywatch
The best free and open-source automated time tracker. Cross-platform, extensible, privacy-focused.
FP8-Emulation-Toolkit
PyTorch extension for emulating FP8 data formats on standard FP32 Xeon/GPU hardware.
smoothquant
[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Quantformer
This is the official pytorch implementation for the paper: *Quantformer: Learning Extremely Low-precision Vision Transformers*.
micronet
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
Paper-Writing-Tips
Paper Writing Tips
Neural-Networks-on-Silicon
This is originally a collection of papers on neural network accelerators. Now it's more like my selection of research on deep learning and computer architecture.
Deep-Learning-Accelerator-SW
NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.
NiuDianNao
A simple cycle-accurate DaDianNao simulator
HolisticTraceAnalysis
A library to analyze PyTorch traces.
TPU-Tensor-Processing-Unit
IC implementation of TPU
gpgpu-sim_distribution
GPGPU-Sim provides a detailed simulation model of contemporary NVIDIA GPUs running CUDA and/or OpenCL workloads. It includes support for features such as TensorCores and CUDA Dynamic Parallelism as well as a performance visualization tool, AerialVisoin, and an integrated energy model, GPUWattch.
Computer-Science-Textbooks
Collect some CS textbooks for learning.
Integrated-Circuit-Textbooks
Collect some IC textbooks for learning.
awesome-model-quantization
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.
AdaptivFloat
Adaptive floating-point based numerical format for resilient deep learning
Deep-Compression-AlexNet
Deep Compression on AlexNet
DynamicViT
[NeurIPS 2021] [T-PAMI] DynamicViT: Efficient Vision Transformers with Dynamic Token Sparsification