Quote:
Originally Posted by Sooga
I've been told SD is basically a sort of average of how far data are from the mean, but it's not even that - that would just be the average of the absolute value differences between the data and mean. So what's the importance of SD? Does it only come into play when you're talking about large samples of normal distributions?
I'm going to approach this question from a slightly different angle.
Instead of thinking about the standard deviation, let's think about the variance. This number is a true average. It's the average of the squared distances from the mean.
But why would we use squared distances?
(1) Absolute values are algebraically annoying. If you remember solving absolute value inequalities, you might remember all of the little "rules" that you have to apply in order to solve them. (If it's of one form, you have an AND condition, but if it's another form it's an OR condition...)
(2) Squaring biases in favor of small distances and against large distances. In other words, it's a weighted average. An error of 1 only contributes 1 to the average, but an error of 2 contributes 4 to the average.
(3) Using absolute values also means that you lose access to some calculus tools, because you can't take the derivative of the absolute value function everywhere.
If you accept this, then at least it would seem plausible that the variance could be interpreted an average of how far data is from the mean. Except that it's not a distance. It's a square distance. If you consider the variables to have units (say, units of length), then the variance is actually a square length.
So we will take the square root at the very end. It's just adjusting the units to be correct. If we want to talk about distances, then we really should be having something with units of distance.
And this is say the standard deviation is kind of like an average distance from the mean.