alphaRGB

alphaRGB

Geek Repo

Company:XiDAIN

Location:Xi'an

Github PK Tool:Github PK Tool

alphaRGB's starred repositories

activitywatch

The best free and open-source automated time tracker. Cross-platform, extensible, privacy-focused.

Language:PythonLicense:MPL-2.0Stargazers:11412Issues:0Issues:0

FP8-Emulation-Toolkit

PyTorch extension for emulating FP8 data formats on standard FP32 Xeon/GPU hardware.

Language:PythonLicense:BSD-3-ClauseStargazers:89Issues:0Issues:0

ppq

PPL Quantization Tool (PPQ) is a powerful offline neural network quantization tool.

Language:PythonLicense:Apache-2.0Stargazers:1462Issues:0Issues:0

smoothquant

[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models

Language:PythonLicense:MITStargazers:1110Issues:0Issues:0

llmtools

Finetuning Large Language Models on One Consumer GPU in Under 4 Bits

Language:PythonStargazers:687Issues:0Issues:0

Quantformer

This is the official pytorch implementation for the paper: *Quantformer: Learning Extremely Low-precision Vision Transformers*.

Language:PythonLicense:Apache-2.0Stargazers:18Issues:0Issues:0

micronet

micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape

Language:PythonLicense:MITStargazers:2203Issues:0Issues:0

Paper-Writing-Tips

Paper Writing Tips

Stargazers:2Issues:0Issues:0

Neural-Networks-on-Silicon

This is originally a collection of papers on neural network accelerators. Now it's more like my selection of research on deep learning and computer architecture.

Stargazers:1811Issues:0Issues:0

GEMM_WMMA

GEMM by WMMA (tensor core)

Language:CudaLicense:Apache-2.0Stargazers:5Issues:0Issues:0

ConvNN

A simple CNN training framework support on CPU and GPU(CUDNN)

Language:C++Stargazers:3Issues:0Issues:0

Deep-Learning-Accelerator-SW

NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.

Language:PythonLicense:NOASSERTIONStargazers:160Issues:0Issues:0
Language:CudaStargazers:91Issues:0Issues:0

NiuDianNao

A simple cycle-accurate DaDianNao simulator

Language:C++License:MITStargazers:10Issues:0Issues:0

HolisticTraceAnalysis

A library to analyze PyTorch traces.

Language:PythonLicense:MITStargazers:254Issues:0Issues:0

nvitop

An interactive NVIDIA-GPU process viewer and beyond, the one-stop solution for GPU process management.

Language:PythonLicense:Apache-2.0Stargazers:4318Issues:0Issues:0

TPU-Tensor-Processing-Unit

IC implementation of TPU

Language:VerilogStargazers:84Issues:0Issues:0

gpgpu-sim_distribution

GPGPU-Sim provides a detailed simulation model of contemporary NVIDIA GPUs running CUDA and/or OpenCL workloads. It includes support for features such as TensorCores and CUDA Dynamic Parallelism as well as a performance visualization tool, AerialVisoin, and an integrated energy model, GPUWattch.

Language:C++License:NOASSERTIONStargazers:1036Issues:0Issues:0

Computer-Science-Textbooks

Collect some CS textbooks for learning.

Stargazers:441Issues:0Issues:0

Integrated-Circuit-Textbooks

Collect some IC textbooks for learning.

Stargazers:88Issues:0Issues:0

viztracer

VizTracer is a low-overhead logging/debugging/profiling tool that can trace and visualize your python code execution.

Language:PythonLicense:Apache-2.0Stargazers:4680Issues:0Issues:0
Language:CudaStargazers:3Issues:0Issues:0
Language:CudaStargazers:45Issues:0Issues:0

goldeneye

GoldenEye is a functional simulator with fault injection capabilities for common and emerging numerical formats, implemented for the PyTorch deep learning framework.

Language:PythonLicense:MITStargazers:22Issues:0Issues:0

PWLQ

Code for our paper at ECCV 2020: Post-Training Piecewise Linear Quantization for Deep Neural Networks

Language:PythonLicense:NOASSERTIONStargazers:66Issues:0Issues:0

awesome-model-quantization

A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.

Stargazers:1733Issues:0Issues:0

AdaptivFloat

Adaptive floating-point based numerical format for resilient deep learning

Language:PythonStargazers:14Issues:0Issues:0

Deep-Compression-AlexNet

Deep Compression on AlexNet

Language:PythonLicense:BSD-2-ClauseStargazers:652Issues:0Issues:0

RobustViT

[NeurIPS 2022] Official PyTorch implementation of Optimizing Relevance Maps of Vision Transformers Improves Robustness. This code allows to finetune the explainability maps of Vision Transformers to enhance robustness.

Language:Jupyter NotebookStargazers:122Issues:0Issues:0

DynamicViT

[NeurIPS 2021] [T-PAMI] DynamicViT: Efficient Vision Transformers with Dynamic Token Sparsification

Language:Jupyter NotebookLicense:MITStargazers:546Issues:0Issues:0