There are 1 repository under bfloat16 topic.
oneAPI Deep Neural Network Library (oneDNN)
Up to 200x Faster Dot Products & Similarity Metrics — for Python, Rust, C, JS, and Swift, supporting f64, f32, f16 real & complex, i8, and bit vectors using SIMD for both AVX2, AVX-512, NEON, SVE, & SVE2 📐
Library for specialized dense and sparse matrix operations, and deep learning primitives.
Half-precision floating point types f16 and bf16 for Rust.
Floating-Point Arithmetic Library for Z80
A LLaMA2-7b chatbot with memory running on CPU, and optimized using smooth quantization, 4-bit quantization or Intel® Extension For PyTorch with bfloat16.
A JAX implementation of stochastic addition.
CUDA/HIP header-only library for low-precision (16 bit, 8 bit) and vectorized GPU kernel development
A Pytorch implementation of stochastic addition.
Comparison of vector element sum using various data types.
Customizable floating point types, with all standard floating point operations implemented from scratch.
Comparison of PageRank algorithm using various datatypes.
Hybridized On-Premise and Cloud (HOPC) Deployment Experimentation with Bfloat16