Intel Corporation's repositories
neural-compressor
SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
intel-extension-for-transformers
⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡
scikit-learn-intelex
Intel(R) Extension for Scikit-learn is a seamless way to speed up your Scikit-learn application
compute-runtime
Intel® Graphics Compute Runtime for oneAPI Level Zero and OpenCL™ Driver
media-driver
Intel Graphics Media Driver to support hardware decode, encode and video processing.
intel-npu-acceleration-library
Intel® NPU Acceleration Library
neural-speed
An innovative library for efficient LLM inference via low-bit quantization
linux-intel-lts
Intel LTS kernel
high-density-scalable-load-balancer
High-Density Scalable Load Balancer(HDSLB) is a high performance Layer-4 load balancer based on DPVS and DPDK(Data Plane Development Kit). It refactors and optimizes the performance of key data plane features, supports 100 Million+ sessions and session sync within LB (load balancer) cluster.
intel-xpu-backend-for-triton
OpenAI Triton backend for Intel® GPUs
cartwheel-ffmpeg
Intel developer staging area for unmerged upstream patch contributions to FFmpeg
llm-on-ray
Pretrain, finetune and serve LLMs on Intel platforms with Ray
cloud-native-ai-pipeline
AI cloud native pipeline for confidential and sustainable computing