Replies: 3 comments 5 replies
-
Yes, and we can even say this expected. There are many issues related in Tensorflow, like INT TFLITE very much slower than FLOAT TFLITE #21698, that
|
Beta Was this translation helpful? Give feedback.
2 replies
-
|
Beta Was this translation helpful? Give feedback.
2 replies
-
Have you tried |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
When I do inference on an Efficientnet model I trained it takes 0.26s for a single image. But when I perform inference on a quantized version of the same Efficientnet model it takes 26 seconds. I thought that because the model is quantized it'd be doing integer operations rather than float operations so should be faster. I wondering if you know what I could do to speed up inference on a quantized model? This is the code Im using:
Beta Was this translation helpful? Give feedback.
All reactions