Hi TI experts,
I was performing frame rate tests by counting the number of frames at the display (measured on the EBD of the display) with the following configuration:
capture->DEI(scale down)->DSP(VLPB)->DEI(scale up)->display.
Where input was in 1080p. Scaling down of the buffer was done x1/4 horizontal and x1/4 vertical total of x1/16 and scaling up was done in x4 horizontal and x4 vertical, total of x16. The frame rate that was measured in this configuration is approximately 59 frames per second.
Now i am using an OMX chain of capture->DSP(VLPB)->Display, where capture is done on 1080p in 422 format, DSP performs only loop-back where the area where the OMX buffers are allocated is cacheable by the DSP and VLPB is compiled with -O3 for maximum performances. To my disappointment the frame rate was deteriorated from 59 frames per second to 30 frames per second, it is obvious that the buffers size at the DSP is making the difference.
My questions are:
1. How can it be? Does the frame rate depends on the buffer size?
2. Since the data buffers are located in shared memory area in order to avoid data copying, only 1 copy occurs and it is copying of input buffer to output buffer, should single HD buffer coping take so many time on a cacheable area?
Thanks,
Gabi