Replies: 2 comments 3 replies
-
If I'm not mistaken, most implementations automatically add both an input quantize and an output dequantize automatically. For clarity, where and how do you apply the quantization? In your machine learning library (TensorFlow, PyTorch), or in nntool? |
Beta Was this translation helpful? Give feedback.
2 replies
-
Your answer is very helpful. I would also like to know if GAP8 supports mixed precision quantization. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, I am trying to deploy ai models on AI-deck to perform tasks. In order to improve the inference speed, I have applied 8-bit post-quantisation to the model, but I want to get the result after dequantization by default, is it possible? How should i implement it?
Beta Was this translation helpful? Give feedback.
All reactions