Our 16-bit quantized model is approximately 10x slower than 8-bit quantized model. We are using TIDL tools 9.2.6.0.
Below are the complete model conversion logs and model inference logs (on VH, debug_level=2)
This thread has been locked.
If you have a related question, please click the "Ask a related question" button in the top right corner. The newly created question will be automatically linked to this question.
Our 16-bit quantized model is approximately 10x slower than 8-bit quantized model. We are using TIDL tools 9.2.6.0.
Below are the complete model conversion logs and model inference logs (on VH, debug_level=2)
Hi,
Adam could you please take initial glance on this issue ?
Lets connect over Webex to discuss this further.
Thanks
Pratik