Answer the question
In order to leave comments, you need to log in
Why MNC only for normal noise?
Why is the use of least squares for approximating time series justified only for those cases where the noise has a normal distribution?
Answer the question
In order to leave comments, you need to log in
The method will work better for the normal distribution of the deviation. it's built into it.
It is easy to show this - you can take some artificial distribution, for example, the deviation is always by a given constant c. Then the function will simply get the offset f`(x) = f(x) + c and approximated by LSM will remain so.
But if the distribution is center-weighted or uniform, then the method will work
Bullshit.
update
Mercury13 is almost right, the only requirement is no odd moments, i.e. even distribution density.
Vasily Melnikov , accordingly, drove the blizzard away, not being able to distinguish a random error (i.e. noise) from a systematic one. It happens.
d0lph1n , having answers like this to his credit , turned out to be an off-topic charlatan . It happens, but it wouldn't be better.
Didn't find what you were looking for?
Ask your questionAsk a Question
731 491 924 answers to any question