You are on page 1of 38

The Seven Pillars of

Statistical Wisdom
Stephen M. Stigler
London School of Hygiene and
Tropical Medicine, January 29, 2015

The Seven Pillars of Statistical Wisdom


A reference to T.E. Lawrence? [of Arabia, whose
1927 memoir was titled Seven Pillars of
Wisdom]
No, rather to Lawrences own source:
Proverbs IX:1
Wisdom hath built her house, she hath hewn
out her seven pillars.
Wisdoms house was built to welcome those
seeking understanding.

Jorge Luis Borgess Chinese


Encyclopedia (1942)
A list divides all animals into one of 14 categories:
Those that belong to the emperor
Embalmed ones
Those that are trained
Suckling pigs
Mermaids
Fabulous ones
Stray dogs
Those that are included in this classification
Those that tremble as if they were mad
Innumerable ones
Those drawn with a very fine camel hair brush
Et cetera
Those that have just broken the flower vase
Those that, at a distance, resemble flies

The Seven Pillars

#1 Aggregation (Sums, Means)


The Combination of Observations: A truly radical
Idea!
Information is gained by discarding information
(discarding individual identification)
Ex. The Mean

#1 Aggregation (Sums, Means)

The Arithmetic Mean, Henry Gellibrand 1635

#1 Aggregation (Sums, Means)

Sumerian 2x3 table, c. 3000 BC, with marginal totals

#1 Aggregation (Sums, Means)


13

22

12

19

16

15

31

31
37

9" 12" 16" 37"


13" 19" 15" 47"
22" 31" 31" 84"
"

47
84

Sumerian 2x3 table, c. 3000 BC, with marginal totals

#1 Aggregation (Sums, Means)

Determining the Lawful Measuring Rod, Kobel 1522

#1 Aggregation (Sums, Means)


More Examples:

Least Squares (Legendre 1805, Gauss 1809)


Index Numbers (Jevons 1863)
Kernel Estimate of a Density (Peirce, 1873)
Nonparametric smoothers

#1 Aggregation (Sums, Means)


12 Mathematicians

#1 Aggregation (Sums, Means)

#2 Information (Root n Rule)


The Ancient Greek Paradox Sorites
(Eubulides of Miletus, circa 350 BC)
The Problem of the Heap:
How do you define a Heap of Sand?
One grain is not a heap.
If you add a grain to a pile of sand that is
not a heap, how then can one grain make
it a heap?

#1 Aggregation (Sums, Means)

#2 Information (Root n Rule)


The Greek Physician Galen (circa 200 AD):
One medical case does not prove
a treatment works.
If you add a single case to a unconvincing
record of cases, how then can that one case
convince?
How can we measure the accumulation of
Information?

#1 Aggregation (Sums, Means)

#2 Information (Root n Rule)


Abraham De Moivre 1730 observed that
the inflection points of the symmetric
Binomial curve for n trials are at a
distance 1 n from the center:
2

#1 Aggregation (Sums, Means)

#2 Information (Root n Rule)


Pierre Simon Laplace
1810 stated a Central
Limit Theorem that
generalized this
(but with a typo!
The square root sign
was missing!)

#1 Aggregation (Sums, Means)

#2 Information (Root n Rule)

G. B. Airy (1861) and Charles S. Peirce (1879)


Extended the idea to more complex settings
(variance component models).

#1 Aggregation (Sums, Means)


#2 Information (Root n Rule)

#3 Likelihood (MLEs,Testing)
The calibration of evidence on a
probability scale.
P- values.
Likelihood based inference, MLEs

#3 Likelihood (MLEs,Testing)
John Arbuthnot, 1711. 82 years with
more Male births than Female.

#3 Likelihood (MLEs,Testing)
The recently discovered
title page of Bayess
1763 Essay:
A Method of Calculating
The Exact Probability of
All Conclusions founded
on Induction.

#1 Aggregation (Sums, Means)


#2 Information (Root n Rule)
#3 Likelihood (MLEs,Testing)

#4 Intercomparison (Percentiles, t-Tests)


Comparisons based upon differences among
the data themselves.
Simplest examples:
1) Percentiles (differences compared to IQR).
2) Students t-test (distance between means
compared to estimated SD of difference).

#4 Intercomparison (Percentiles, t-Tests)


More complicated examples:

1) ANOVA with structured data (blocks,


split plots, additive models, etc).
2) The Bootstrap

#1 Aggregation (Sums, Means)


#2 Information (Root n Rule)
#3 Likelihood(MLEs,Testing)
#4 Intercomparison (Percentiles, t-Tests)

#5 Regression (Multivariate Inference


and Bayes Theorem)
Tall parents on average produce somewhat
shorter children than themselves;
tall children on average have somewhat
shorter parents than themselves.

#5 Regression (Multivariate Inference and Bayes Theorem)

Two marginal
distributions,
Two conditional
expectations;
Bayes Theorem!

#5 Regression (Multivariate Inference and Bayes Theorem)

B level
A

A level

B
A

Galton, 1885: Two regression lines


(two lines of conditional expectation).

#1 Aggregation (Sums, Means)


#2 Information (Root n Rule)
#3 Likelihood(MLEs,Testing)
#4 Intercomparison (Percentiles, t-Tests)
#5 Regression (Multivariate Inference and Bayes Theorem)

#6 Design (ANOVA, Design-Based


Inference)
Examples:
Book of Daniel (Old Testament)
Avicenna (or Ibn Sina, circa 1000 AD)
Sample surveys

#6 Design (ANOVA, Design-Based


Inference)

Random assignment in pairs (Peirce 1880s)


Induction: Reasoning from a sample taken
at random to the whole lot sampled

#6 Design (ANOVA, Design-Based


Inference)
W. S. Jevons 1873: One of the most requisite
precautions in experimentation is to vary only
one circumstance at a time, and to maintain
all other circumstances rigidly unchanged.
R. A. Fisher 1926: No aphorism is more
frequently repeated in connection with
field trials, than that we must ask Nature
few questions, or, ideally, one question,
at a time. this view is wholly mistaken.

#6 Design (ANOVA, Design-Based


Inference)

Random Latin Square designs (Fisher 1930s)


Spherical symmetry approximated by
random design, validating tests!

#1 Aggregation (Sums, Means)


#2 Information (Root n Rule)
#3 Likelihood (MLEs,Testing)
#4 Intercomparison (Percentiles, t-Tests)
#5 Regression (Multivariate Inference and Bayes Theorem)
#6 Design (ANOVA, Design-Based Inference)

#7 Residuals (Models and Model


Comparison)

#7 Residuals (Models and Model Comparison)

John Herschel, 1831:


Complicated
phenomenamay be
simplified by
subducting the effect of
known causes, leaving
a residual phenomenon to be
explained.
It is by this processthat science is chiefly promoted.

#7 Residuals (Models and Model Comparison)

John Stuart Mill, 1843, in A System of Logic, called


this The Method of Residues:
Of all the methods of investigating laws of
nature, this is the most fertile in unexpected
results.
In statistics:
Residual plots as diagnostics
Comparison of nested models
Testing significance of regression coefficients
Parametric modelling
Partial Likelihood (Cox models)

#1 Aggregation (Sums, Means)


#2 Information (Root n Rule)
#3 Likelihood (MLEs,Testing)
#4 Intercomparison (Percentiles, t-Tests)
#5 Regression (Multivariate Inference and Bayes Theorem)
#6 Design (ANOVA, Design-Based Inference)
#7 Residuals (Models and Model Comparison)

#1 Targeted reduction/compression of data


#2 Diminishing value of more data
#3 Putting a probability measure to inferences
#4 Doing this based upon internal data variation
#5 Different perspectives give different answers
#6 The Essential role of planning
#7 How to explore in nested families of models

#1 Aggregation (Sums, Means)


#2 Information (Root n Rule)
#3 Likelihood (MLEs,Testing)
#4 Intercomparison (Percentiles, t-Tests)
#5 Regression (Multivariate Inference and Bayes Theorem)
#6 Design (ANOVA, Design-Based Inference)
#7 Residuals (Models and Model Comparison)

Revolutionary ideas:
#1 Gain information by suppressing information
W. S. Jevons replying to critics:
Were a complete explanation of each fluctuation
thus necessary, not only would all inquiry into
this subject be hopeless, but the whole of the
statistical and social sciences, so far as they
depend upon numerical facts, would have
to be abandoned.

#1 Aggregation (Sums, Means)


#2 Information (Root n Rule)
#3 Likelihood (MLEs,Testing)
#4 Intercomparison (Percentiles, t-Tests)
#5 Regression (Multivariate Inference and Bayes Theorem)
#6 Design (ANOVA, Design-Based Inference)
#7 Residuals (Models and Model Comparison)

More Revolutionary Ideas:


#2 All data equally good but added data less valuable
#3 Probability as a measure for inference
#4 Measure accuracy with no exterior standard
#5 Euclid proportionality wrong in all interesting science
#6, #7 Major change to scientific experimentation

#1 Aggregation (Sums, Means)


#2 Information (Root n Rule)
#3 Likelihood (MLEs,Testing)
#4 Intercomparison (Percentiles, t-Tests)
#5 Regression (Multivariate Inference and Bayes Theorem)
#6 Design (ANOVA, Design-Based Inference)
#7 Residuals (Models and Model Comparison)

Features or bugs?
#1 Treat people as mere statistics
#2 Is Big Data good based on size alone?
#3#4 Significance tests: Problem or solution?
#5#6#7 Criticism of regression models, fitting, selection
E. B. Wilson 1927: It is largely because of lack of
knowledge of what statistics is that the person
untrained in it trusts himself with a tool quite as
dangerous as any he may pick out from the whole
armamentarium of scientific methodology.

#1 Aggregation (Sums, Means)


#2 Information (Root n Rule)
#3 Likelihood (MLEs,Testing)
#4 Intercomparison (Percentiles, t-Tests)
#5 Regression (Multivariate Inference and Bayes Theorem)
#6 Design (ANOVA, Design-Based Inference)
#7 Residuals (Models and Model Comparison)

These Seven Pillars are not mathematics


and are not Computer Science. They do
centrally constitute the important core ideas
underlying the Science of Statistics.

The Seven Pillars of Statistical Wisdom


Proverbs IX:1
Wisdom has built her house,
she has hewn her seven pillars.
Wisdom has built her house,
the seven have set its foundations.
Refers to the Seven Sages of the seven ancient
(before the great flood) cities of Mesopotamia

Three Local Pillars of Wisdom

Major Greenwood, Austin Bradford Hill, Peter Armitage

Seven (+) Sages.

This content downloaded from 76.192.186.43 on Fri, 29 Nov 2013 12:53:18 PM


All use subject to JSTOR Terms and Conditions

You might also like