There are 17 repositories under inference-engine topic.
FEDML - The unified and scalable ML library for large-scale distributed training, model serving, and federated learning. FEDML Launch, a cross-cloud scheduler, further enables running any AI jobs on any GPU cloud or on-premise cluster. Built on this library, FEDML Nexus AI (https://fedml.ai) is your generative AI platform at scale.
Rule engine implementation in Golang
带你从零实现一个高性能的深度学习推理库,支持大模型 llama2 、Unet、Yolov5、Resnet等模型的推理。Implement a high-performance deep learning inference library step by step
FeatherCNN is a high performance inference engine for convolutional neural networks.
OneDiff: An out-of-the-box acceleration library for diffusion models.
Paddle.js is a web project for Baidu PaddlePaddle, which is an open source deep learning framework running in the browser. Paddle.js can either load a pre-trained model, or transforming a model from paddle-hub with model transforming tools provided by Paddle.js. It could run in every browser with WebGL/WebGPU/WebAssembly supported. It could also run in Baidu Smartprogram and WX miniprogram.
PygmalionAI's large-scale inference engine
High performance Cross-platform Inference-engine, you could run Anakin on x86-cpu,arm, nv-gpu, amd-gpu,bitmain and cambricon devices.
A common base representation of python source code for pylint and other projects
Python Computer Vision & Video Analytics Framework With Batteries Included
docs for search system and ai infra
The Qualcomm® AI Hub Models are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
This is a repository for an object detection inference API using the Tensorflow framework.
A quick view of high-performance convolution neural networks (CNNs) inference engines on mobile devices.
TinyTensor is a tool for running already trained NN (Neural Network) models to be able to use them for inference of various tasks such as image classification, semantic segmentation, etc.
Repository for OpenVINO's extra modules
Large scale recommender system inference Microservices and APIs (Dubbo 、gRPC and REST ) with Golang.
Deep computer-vision algorithms for the Processing framework.
Neural network inference template for real-time cricital audio environments - presented at ADC23
A Machine Learning System for Data Enrichment.
An optimized neural network operator library for chips base on Xuantie CPU.
This is a repository for a No-Code object detection inference API using the OpenVINO. It's supported on both Windows and Linux Operating systems.