Replies: 3 comments 5 replies
-
Yes, and we can even say this expected. There are many issues related in Tensorflow, like INT TFLITE very much slower than FLOAT TFLITE #21698, that
|
Beta Was this translation helpful? Give feedback.
2 replies
-
|
Beta Was this translation helpful? Give feedback.
2 replies
-
Have you tried |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
When I do inference on an Efficientnet model I trained it takes 0.26s for a single image. But when I perform inference on a quantized version of the same Efficientnet model it takes 26 seconds. I thought that because the model is quantized it'd be doing integer operations rather than float operations so should be faster. I wondering if you know what I could do to speed up inference on a quantized model? This is the code Im using:
Beta Was this translation helpful? Give feedback.
All reactions