jiqing-feng's repositories
accelerate
🚀 A simple way to train and use PyTorch models with multi-GPU, TPU, mixed-precision
bitsandbytes
Accessible large language models via k-bit quantization for PyTorch.
ClipBERT
[CVPR 2021 Best Student Paper Honorable Mention, Oral] Official PyTorch code for ClipBERT, an efficient framework for end-to-end learning on image-text and video-text tasks.
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
diffusers
🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch
Diffusion-MU-Attack
The official implementation of the paper "To Generate or Not? Safety-Driven Unlearned Diffusion Models Are Still Easy To Generate Unsafe Images ... For Now". This work introduces one fast and efficient attack methods to generate toxic content for safety-driven diffusion models.
FlexFlow
A distributed deep learning framework.
GEAR
GEAR: An Efficient KV Cache Compression Recipefor Near-Lossless Generative Inference of LLM
intel-extension-for-transformers
Extending Hugging Face transformers APIs for Transformer-based models and improve the productivity of inference deployment. With extremely compressed models, the toolkit can greatly improve the inference efficiency on Intel platforms.
lm-evaluation-harness
A framework for few-shot evaluation of autoregressive language models.
neural-compressor
Intel® Neural Compressor (formerly known as Intel® Low Precision Optimization Tool), targeting to provide unified APIs for network compression technologies, such as low precision quantization, sparsity, pruning, knowledge distillation, across different deep learning frameworks to pursue optimal inference performance.
optimum
🏎️ Accelerate training and inference of 🤗 Transformers with easy to use hardware optimization tools
optimum-intel
Accelerate inference of 🤗 Transformers with Intel optimization tools
peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
pytorch_geometric
Graph Neural Network Library for PyTorch
tau
Pipeline Parallelism for PyTorch
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
models
Intel® AI Reference Models: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors and Intel® Data Center GPUs
ProtST
Camera-ready repo for ProtST
q-diffusion
[ICCV 2023] Q-Diffusion: Quantizing Diffusion Models.
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs