Hi,
I am currently working on C6678 multicore DSP and made a small benchmark test. What I tested is the time it takes to multiply two 3x3 sized matrices for both in debug mode and in optimization mode (-O3). I measured the time using RTOS analyzer exec graph. I multiplied the two matrices 1000 times and calculated the mean CPU time for it.
So here are the results of average time:
Debug Mode, Float(REAL32) : 1.251us (micro seconds)
Optimization Mode, Float(REAL32) : 0.231us
Debug Mode, Double(REAL64) : 1.262us
Optimization Mode, Double(REAL64) : 0.228us
Here is my question: Are not double operations supposed to take longer CPU time than float operations? Why in optimization mode double type definition shows better performance?
Thank you.