Nimrod Rak's repositories
server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
Language:PythonBSD-3-Clause000
tempo
Grafana Tempo is a high volume, minimal dependency distributed tracing backend.
Language:GoAGPL-3.0000
TensorRT
NVIDIA® TensorRT™, an SDK for high-performance deep learning inference, includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for inference applications.
Language:C++Apache-2.0000
tiny-tensorrt
Deploy your model with TensorRT quickly. 快速使用TensorRT来部署模型
Language:C++000
torch2trt
An easy to use PyTorch to TensorRT converter
Language:PythonMIT000
website
Kubernetes website and documentation repo:
Language:HTMLCC-BY-4.0000