Can someone please explain the meaning of the "accuracy" value given back in the SW API's for most sensor measurements for BHI160 and BMM150. I undertstand the algorithms train themselves to try and minimise errors and maximise accuracy.
We are currently reseting the whole chip complex and re'initialising between taking measurements. I think this is probably wrong; meaning we would always start sampling from a low accuracy point. How long does sensor complex have to be running to reach a state of best accuracy?
I have seen a definition was posted here for one of the other sensors.
Is this defined anywhere for the BHI160 & BMM150?