Part Number: AM62A7
Tool/software:
Greetings!
I am facing a strange accuracy issue when offloading the execution of my model to the hardware accelerators. I’ve checked the troubleshooting guides and played around with the accuracy_level and tensor_bits parameters of the model, but that had a marginal positive impact on the results. From what I have managed to identify so far, the problem seem to be tightly related to either the number of inputs the model is supposed to process, or the order in which the inputs are processed.
To exemplify my problem, let’s consider my model, which is designed to process an input with 4 metadata features and give it a label. The recurring issue I’ve been facing is that, given a batch of X inputs to be processed on the hardware accelerator, only the first input is consistently correctly labeled by the model.

The figure below illustrates my problem: if I feed a batch of 5 inputs to the CPU, the accuracy rarely goes below 98%. If I feed the same batch to the offloaded model, only the first input is concistently correctly labeled, while the remaining 4 inputs face a drastic drop in accuracy. As highlighted in green, the first input of the offloaded model closely relates to the values inferred in the CPU for the same data set. This behavior is reproducible with other batch values as well, like 20, 100, etc. As can be expected, when the batch number is 1, the accuracy is always on par with the CPU.

Do you happen to have any insights on what could be happening? I’ve checked posts from other users, but their problems didn’t seem to have anything in common with what I am experiencing.
Thank you for your time!
Best regards,
Giann.
