Tool/software:
In our image classification pipeline, the TI model-zoo models require fp32 input.
We are currently converting UINT8 image data to fp32 on the ARM processor, normalizing it, and then passing it to the Deep Learning Accelerator.
If the Deep Learning Accelerator were capable of performing the normalization, we could bypass the fp32 conversion and directly input the UINT8 images, which should reduce latency.
Is this feasible?