I have two questions regarding the statement that the sensor linearity error of the DRV5055 is approximately ±1%.
1.This states that linearity error is defined as the difference in sensitivity between any two positive B values and any two negative B values. Will the error change if used only in the positive range? Or does the error change when used only in the negative range?
2. To consider the worst case of error, suppose two arbitrary B values are 100mT and 101mT. Assume that the error is 0% at 100mT and -1% at 101mT. The sensitivity of the sensor is 7.5mV/mT.
In that case, is the calculation correct that the output voltage for 100mT is 750mV with 0% error, and 749.925mV for 101mT with -1% error? Is it possible that the output voltage decreases even if the magnetic flux increases?
Or is it correct to think that the output voltage of 101mT is 757.425mV, which is the difference between the two B values, 1mT*7.5*0.99, added to 750mV (100mT)?