Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why does quantified model run slower? #2

Open
oroelipas opened this issue Mar 5, 2020 · 4 comments
Open

Why does quantified model run slower? #2

oroelipas opened this issue Mar 5, 2020 · 4 comments

Comments

@oroelipas
Copy link

I know these models are designed to run in Coral USB but should't also it run faster in a PC?

It takes about 1.15 seconds to run a tiny-yolov3.tflite in my coumputer but arround 15 seconds to run the quant_coco-tiny-v3-relu.tflite

Is this normal behaviour?
will it run faster than 1.15s in Coral USB TPU when I buy it?

Thanks for answering!

@guichristmann
Copy link
Owner

I don't know the reason, but I also got a bad performance when running the quantized model on the desktop CPU. However, it runs fine/as expected with the Edge TPU.

@parthjdoshi
Copy link

@guichristmann What kind of inference speeds did you get with YOLOv3-tiny? In terms of FPS?

@AkkiSony
Copy link

@oroelipas @parthjdoshi @guichristmann Hi. I could convert my yolov3 model to quantized tflite model. But when I try to run the inference.py script. I do not see any detection happening on the input image.
Were you able to get it doneusing this platform? Did you make any changes in the script?
Please do help me get this run. Thank you in advance guys.

@lukqw
Copy link

lukqw commented Oct 27, 2021

@parthjdoshi I get around 16-17 fps with a yolov3-tiny model with relu as the activation function on device with edge tpu

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants