Ivan Yashchuk's starred repositories
mesh-transformer-jax
Model parallel transformers in JAX and Haiku
BigDL
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, ModelScope, etc.
Essentials-of-Compilation
A book about compiling Racket and Python to x86-64 assembly
hlb-CIFAR10
Train to 94% on CIFAR-10 in <6.3 seconds on a single A100. Or ~95.79% in ~110 seconds (or less!)
resource-stream
CUDA related news and material links
bigscience
Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.
programming-language-research
✨ Programming Language Research, Applied PLT & Compilers
mlir-tutorial
MLIR For Beginners tutorial
illustrated-machine-learning.github.io
Website containing illustrations about Machine Learning theory!
pythonflow
:snake: Dataflow programming for python.
Tranformers_Are_What_You_Dont_Need
The best repository showing why transformers don’t work in time series forecasting and showcasing the best SOTA non transformer models.
auto-walrus
Automatically use the awesome walrus operator
pysr_scaling_laws
You should use PySR to find scaling laws. Here's an example.
CassetteOverlay.jl
An experimental simple method overlay mechanism for Julia
advanced-mpi
Advanced MPI and parallel I/O