who who who's repositories
tensorflow
An Open Source Machine Learning Framework for Everyone
composer
Supercharge Your Model Training
examples
Fast and flexible reference benchmarks
flash-attention
Fast and memory-efficient exact attention
FlexGen
Running large language models on a single GPU for throughput-oriented scenarios.
jax
Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
keras-io
Keras documentation, hosted live at keras.io
llama
Inference code for LLaMA models
llm-foundry
LLM training code for MosaicML foundation models
Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
MIOpen
AMD's Machine Intelligence Library
nccl-rccl-parser
Tool to run rccl-tests/nccl-tests based on from an application
pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
triton
Development repository for the Triton language and compiler
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs