There are 0 repository under inference-library topic.
Describing How to Enable OpenVINO Execution Provider for ONNX Runtime
Latte is a convolutional neural network (CNN) inference engine written in C++ and uses AVX to vectorize operations. The engine runs on Windows 10, Linux and macOS Sierra.
Rust library managing long conversations with any LLM
Unified JavaScript API for scoring via various DL framework
Experimental Python implementation of OpenVINO Inference Engine (very slow, limited functionality). All codes are written in Python. Easy to read and modify.
NodeJS binding for Menoh DNN inference library