I'm trying to understand the various configurations of the object detection DPC and how to utilize the DSP core.
Reading the documentation and the source code for the demos, it seems like there are two methods in the examples. The AOP example uses the HWAs (objdethwa) for all of the steps of object detection, and the ARM core to handle control and callbacks. The ISK demo uses the DSP (objdetdsp) for the entire chain after RangeProc.
But reading the documentation, it seems like there's an option to run HWA accelerated algorithms but dispatched from the DSP. I'm drawing this conclusion from the different DPU docs each saying the "HWA" version will run on the R4F or the DSP cores. Am I understanding that correctly?
Is there a code example of running the accelerated algorithms from the DSP core? Which documentation would help me in understanding the software development model used for building/linking combined MSS+DSS programs?
(What am I trying to do? Send back the largest, densest pointcloud possible to the host over UART. After switching to external 2Mbps UART and compressing coordinates, my bottleneck now is simply inter-frame processing time in the AOP demo. I want to try moving some or all processing to the DSP, as it has a faster core clock.)