monoloxo / tensorrt-yolov5

Run tensorrt yolov5 on Jetson devices, supports yolov5s, yolov5m, yolov5l, yolov5x.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

OpenJetson

http://openjetson.com/

yolov5

The Pytorch implementation is ultralytics/yolov5.

Currently, we support yolov5 v1.0(yolov5s only), v2.0, v3.0 and v3.1.

Config

  • Choose the model s/m/l/x by NET macro in yolov5.cpp
  • Input shape defined in yololayer.h
  • Number of classes defined in yololayer.h, DO NOT FORGET TO ADAPT THIS, If using your own model
  • FP16/FP32 can be selected by the macro in yolov5.cpp
  • GPU id can be selected by the macro in yolov5.cpp
  • NMS thresh in yolov5.cpp
  • BBox confidence thresh in yolov5.cpp
  • Batch size in yolov5.cpp

How to Run, yolov5s as example

1. generate yolov5s.wts from pytorch with yolov5s.pt, or download .wts from model zoo

git clone https://github.com/wang-xinyu/tensorrtx.git
git clone https://github.com/ultralytics/yolov5.git
// download its weights 'yolov5s.pt'
// copy tensorrtx/yolov5/gen_wts.py into ultralytics/yolov5
// ensure the file name is yolov5s.pt and yolov5s.wts in gen_wts.py
// go to ultralytics/yolov5
python gen_wts.py
// a file 'yolov5s.wts' will be generated.

2. build tensorrtx/yolov5 and run

// put yolov5s.wts into tensorrtx/yolov5
// go to tensorrtx/yolov5
// ensure the macro NET in yolov5.cpp is s
mkdir build
cd build
cmake ..
make
sudo ./yolov5 -s         // serialize model to plan file i.e. 'yolov5s.engine'
sudo ./yolov5 -v         // deserialize plan file and run inference with camera or video.

demo

About

Run tensorrt yolov5 on Jetson devices, supports yolov5s, yolov5m, yolov5l, yolov5x.


Languages

Language:C++ 73.1%Language:Python 13.2%Language:Cuda 12.5%Language:CMake 1.2%