The GitHub Archive Program & Arctic Code Vault
The repository for arewegameyet.rs
Bolt is a deep learning library with high performance and heterogeneous flexibility.
Chisel 3: A Modern Hardware Design Language
A tool which profiles OpenCL devices to find their peak capacities
The fastest font renderer in the world, written in pure rust.
MACE is a deep learning inference framework optimized for mobile heterogeneous computing platforms.
Push-button installer of macOS Catalina, Mojave, and High Sierra guests in Virtualbox for Windows, Linux, and macOS
MNN is a blazing fast, lightweight deep learning framework, battle-tested by business-critical use cases in Alibaba
ncnn is a high-performance neural network inference framework optimized for the mobile platform
oneAPI Deep Neural Network Library (oneDNN)
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
Multi-platform high performance deep learning inference engine (『飞桨』多平台高性能深度学习预测引擎）
RISC-V Assembly code assembler package for Python.
The repository for rust-gamedev.github.io
Tengine is a lite, high performance, modular inference engine for embedded device
An Open Source Machine Learning Framework for Everyone
TNN: developed by Tencent Youtu Lab and Guangying Lab, a lightweight and high-performance deep learning framework for mobile inference. TNN is distinguished by several outstanding features, including its cross-platform capability, high performance, model compression and code pruning. Based on ncnn and Rapidnet, TNN further strengthens the support and performance optimization for mobile devices, and also draws on the advantages of good extensibility and high performance from existed open source efforts. TNN has been deployed in multiple Apps from Tencent, such as Mobile QQ, Weishi, Pitu, etc. Contributions are welcome to work in collaborative with us and make TNN a better framework. TNN：由腾讯优图实验室和光影实验室协同打造，移动端高性能、轻量级推理框架，同时拥有跨平台、高性能、模型压缩、代码裁剪等众多突出优势。TNN框架在原有Rapidnet、ncnn框架的基础上进一步加强了移动端设备的支持以及性能优化，同时也借鉴了业界主流开源框架高性能和良好拓展性的优点。目前TNN已经在手Q、微视、P图等应用中落地，欢迎大家参与协同共建，促进TNN推理框架进一步完善。
Open deep learning compiler stack for cpu, gpu and specialized accelerators
Native WebGPU implementation based on gfx-hal
Rust bindings to wgpu native library