wobjtushisui / model_analyzer

Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Server models.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

License

Triton Model Analyzer

NOTE: You are currently on main branch. This branch is tracking development for the new version of Model Analyzer and is not currently usable.

Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Server models. These reports will help the user better understand the trade-offs in different configurations and choose a configuration that maximizes the performance of Triton Inference Server.

Requirements

Triton Model Analyzer supports all the GPUs supported by the DCGM library. See DCGM Supported GPUs for more information.

Documentation

Reporting problems, asking questions

We appreciate any feedback, questions or bug reporting regarding this project. When help with code is needed, follow the process outlined in the Stack Overflow (https://stackoverflow.com/help/mcve) document. Ensure posted examples are:

  • minimal – use as little code as possible that still produces the same problem

  • complete – provide all parts needed to reproduce the problem. Check if you can strip external dependency and still show the problem. The less time we spend on reproducing problems the more time we have to fix it

  • verifiable – test the code you're about to provide to make sure it reproduces the problem. Remove all other problems that are not related to your request/question.

About

Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Server models.

License:Apache License 2.0


Languages

Language:Python 91.8%Language:Shell 7.8%Language:Smarty 0.2%Language:Dockerfile 0.2%