AlexeyAB / yolo2_light

Light version of convolutional neural network Yolo v3 & v2 for objects detection with a minimum of dependencies (INT8-inference, BIT1-XNOR-inference)

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

is that possible to use full INT8 during tiny-yolo-v3

jasonwu1977 opened this issue · comments

Hi @AlexeyAB
I would like to know is that possible to use full INT8 during all yolo-light2

  1. after pre-processing layer
    is that possible to convert it to INT8 before entering the first layer
  2. can i use INT8/INT16 bias
  3. can I save output to 8 bit before next layer (of course befoer yolo layer it still float32)

@AlexeyAB
Answer for myself
All 3 questions are yes, and i have tested with CPU

I want to run quantization with pure CPU mode.
And When I change the Makefile to let yolo2-light running on CPU mode.
the mAP drops to very low, after I trace the code, i noticed the quantization of CUDNN mode is different with running GPU mode (some layer don't quantize in CUDNN mode)

Now the question is, on yolov2 or yolov3, on CUDNN examples, the size=1 & stride=2 layer doesn't do the quantize, but is it possible to do quantize with CPU mode?
And if i want to change the CPU to support all convolution layer quantization, where should i moidfy?

@jasonwu1977 hi Jason, I met similar issue as you, could you can talk it more about it through my qq 494529371? big thanks.

Have you sloved it out yet? or anyone else know?