Part Number: J721EXCPXEVM
Hi:
My model is a key point detection model (using the sigmoid activation function at the end). I used the QAT tool provided by Ti to quantize my model to 16 bits. The accuracy of the 16bit quantized model did not decrease, which is great.
In order to make the model inference faster on TDA4, QAT is used to quantify the model into 8-bit, but the accuracy of the 8-bit model is reduced by about 10% compared to the float model, which is unacceptable.
I have taken the suggestions in "ti_dl/docs/user_guide_html/md_tidl_fsg_quantization.html".
How can the accuracy of the 8-bit model not decrease?
The sdk version is 7.2.
Best regards,
Gong