Hello,
I have adapted an example VI that implements a Kalman filter in a control & simulation loop. By default, the step size for the control & simulation loop is 0.02 seconds, but when I run the program with this setting, the
Kalman filter does not behave correctly (it does not converge to the value I expect it to converge to). I tried reducing the step size to 0.00005 seconds, and it appears that the Kalman filter works better, but is this an untypically small step size? How can I tell that I have selected an appropriate step size (esp. for a Kalman filter)? Is a rule of thumb to pick as large of a step size as possible without affecting the functionality of your program (such that the step size is not so small as to reduce performance)? Thank you.