This is the second part of a series of writings talking about the difficulties of probabilistic reasoning; here, we zoom in the problem of applying probability to actual data, examining the paradigms and techniques of the field of statistics.
Our focus is, similarly to the first article, on the hidden assumptions people make, the effects these assumptions have on the validity of their inferences, and the absence of perfect solutions — in short, why statistics is difficult. Along the way, we’ll be introducing lots of statistical frameworks, techniques, and models, generally at a relatively rigorous level. You don’t necessarily have to follow the more rigorous aspects of each argument and derivation to get the gist of it, but a familiarity with calculus helps immensely. Useful properties of some common distributions are given in part A of the Appendices, while the most technical and/or tedious derivations are stowed away in part B.
Table of Contents
Part 1: Probability is Difficult
Part 2: Statistics is Difficult
Appendix: Uncertain Appendices
In the previous article, we covered the basic interpretations of probability:
Now, we’ll put probability to the test by figuring out how to use it to understand the world. That it is useful is clear, for it is used all the time — to determine the reliability of industrial processes, to predict the fluctuations of the stock market, to verify the accuracy of measurements, and, in general, to help us make informed decisions. Hence, we must understand why, how, and when it is useful, and how to use it correctly.
To paraphrase Bandyopadhyay’s Philosophy of Statistics, there are at least four different motivations for the use of statistical techniques:
Obviously, these questions are very tightly connected to one another, but each of them provokes us to probe the data in different ways, to ask different follow-up questions, and to interpret statements about the data in different ways; as such, they segregate themselves into different interpretations of statistical inference.
At the same time, there are different schools of statistical inference, which generally (but not always) hold fast to one of these interpretations. The distinction I’m making between schools and interpretations is a rather subtle one, but will be very important; to put it simply, a school is an established body of interlocking techniques of statistical inference, whereas an interpretation is an assigned purpose for those techniques.
A short synopsis of the main schools of statistical inference: