There are 0 repository under int8-inference topic.
Real-time portrait segmentation for mobile devices
Generate a quantization parameter file for ncnn framework int8 inference
BEVFormer inference on TensorRT, including INT8 Quantization and Custom TensorRT Plugins (float/half/half2/int8).
Quantization Aware Training
将端上模型部署过程中,常见的问题以及解决办法记录并汇总,希望能给其他人带来一点帮助。
TensorRT Int8 Python version sample. TensorRT Int8 Python 实现例子。TensorRT Int8 Pythonの例です
GPT-J 6B inference on TensorRT with INT-8 precision
VB.NET api wrapper for llm-inference chatllm.cpp
Generating tensorrt model using onnx
C# api wrapper for llm-inference chatllm.cpp
it has support for openvino converted model of yolov7-int.xml ,yolov7x,
Compressed CNNs for airplane classification in satellite images (APoZ-based parameter pruning, INT8 weight quantization)