Professional Documents
Culture Documents
THEORY
OF
STATISTICS
1. Distribution Theory
2. Estimation
3. Tests of Hypotheses
"
1958-1959
North Carolina state College
Notes recorded by R. C.. Taeuber
Institute of Statistics
Mimeo Series No. 241
September, 1959
'J
"
- ii TABLE OF CONTENTS
I. PRELIl1INARIES
Concepts of set theory
Probability measure
Distribution function (Th. 1)
Density funct ion (def. 8)
o
0
"
III.
1
3
5
8
10
10
12
12
15
CHARACTERISTIC FUNCTIONS
0
0
0
0
0
0
0
0
0
Standard distributions
Expectation and moments
Tchebycheff 's Inequality (Th. e + corollary)
Characteristic functions
Cumulant generating functions (def. 17)
Inversion theorem for characteristic functions
Central Limit Theorem (Tho 14)
20
" 21
.. 24
e 25
28
0
I)
(Tho 12)
C)
30
33
36
36
36
CONVERGENCE
Convergence of distributions 0 0 0 0 0 e 0 39
Convergence in probability (def. 18) " 41
Weak law of large numbers (Th. 16)
41
A Convergence Theorem (Th~ 18) 0 0 0 0 0 44
Convergence almost everywhere (def. 20)
e
49
Strong law of large numbers (Th. 20)
51
Other types of convergence 0 0 0 0 " 0 0 53
0
IV.
"
"
POINT ESTIMATION
Preliminaries
Ways to formulate the problem of obtaining best estimates
Restricting the class of estimates
optimum properties in the large
Dealing only with large sampla (asymptotic) properties
Methods of Estimation
I'1ethod of moments e e
"
Method of least squares 0
Gauss-Markov Theorem (Th. 21) ,
"
Maximum Likelihood Estimates
Single parameter estimation (Th. 22)
Multiparameter estimation (Tho 23) 0 0 0
0
III
I)
ill
55
56
58
60
It
61
61
,.
61
64
65
70
-t
..
..
Examples ".. .
Minimax estimation
c
Wolfowitzts minimum distance estimation
0
0
DO'
V.
~EST1NG
..
..
"
fi
..
..
..
76
80
85
86
88
92
95
98
102
102
103
103
104
10,
108
110
Basic concepts
113
Distribution free tests
U5
One sample tests
o. 117
Probability transformation
~ 121
Tests for one sample problems
121
Convergence of sample distribution function (Tho 31)
124
Approaches to combining probabilities o. 129
Two sample tests 0 0 41 0 0 0 133
Pitman I s theorem on asymptotic relative efficiency (Tho 32) 140
k-sample tests 0 0 .. 0 0 0 0 ,. 0 ,. 146
0
....
.'
....
..
VI.
0
o Q o "
0
41
....
'"
..
..
..
41
148
149
156
161
163
168
169
173
175
182
186
186
187
191
191
192
VIII.'
I{(SCELLANEOUS~
193
0
194
197
CONFlDENCE REGIONS
202
.,
204
206
I:
Preliminaries:
Set~
Def",
A collection of points in
1/ 8J.
lie
(Sr e
in S:t
it' 81 is contained in 8
1 '" 8 2
~S2
c.l
8:l.
and 52'
82 + ~
I:
8 251
There is also the obvious extension of these definitions of set addition and multiplication to 3 or more sets.
CD
~ 18n
II;l
&:
CD
-nn'J 18n
s*
Proof':
denote
Let
X. 8
iGeo x
!l1is
s.
J 81J
therefore x e
an element of n
82
'*
~ ~ 1
e 8 2*
Sr*
8 2*~ x
B ~
* 82*
I;
=- ~
x e 81* 8 2*
* and x
:l 6
~ x_~
e 8 2*
andxJS2
. 7
x /.
(f\
::::::::r
(81 + S2)*'
+ 82 )
Exercise
2/
Exercise
1/
In R define 81 .. {x,y: x 2 + y2 ~
2
1}
82
{x,;r
{x,y:
z O}
Represent graphically ~ + 8 2,
Def. 2/ If ~ c. 82 c 8
S:J.* 8 2"
S3S251~ 81 8 2 - SlS;"
co
We
def'ine~
lim Sn'" ~ Sn
n-l
n~(X)
CD
We
define~
lim Sn
n~tO
fr.B
n .1 n
hi
{x,;r: txt
Ans~
(b)
8n [x,p1'g
Ixl
<:
1 +~,
lyl ~l
may be represented
1Y/< 1 + ~
11'1 <1
can be represented
-.1-
3/ Borel
Sets -
The word multiplication oould be deleted, since multiplication can be performed by complementation, eog4l:
(.tIL * + S
-~
Def.
4/
*)*
(s*)*
S. S
-J. 2
11 for
Examples: -
- in!l:L
C?'
~(S).
A(S). !f(X) dx
2Y
f:
dx
~ef3
5/
3/ P(I1c) '" 1
C;;
P() .. 0
Exo
6/
Lemma 2/ P (~ .;. 8 2 +
Problem 1:
Frovc lcr.una 2.
0)
~ P(~) + P(S2) +
1 oeo
til
R *;
k
9J
1\
t:t
R
k
Proof:
case I Define:
~ C 8 C 8,
~.
Sf
st
3
82 - ~
S3 - 52
etc.
CD
n-l
CD
Sn
n-l n
CD
P ( 11m Sn)
n~CD
(~
8n )
-1
00
(~ S~)
n-l
P (~) + P (S2-51 ) + P (8 - 8 2 )
3
P(~,>
..P (8 ) + P(S2)
1
- P(S2) + P(S3)
etc.
P (Sn) after n steps
lim P(Sn)
n~
Case 2:
~::>
8 2 ,:) 8
rex)
peS) on
defined by
F(x.)
pc
00 I
x).
dolo
Ii. there
is a point function
1/ For Xl < x 2 we
The interval (.
From exercise
00" ~)
6 we
PC:;) ~ F(X2 )
C "'he interval (.
have that
P(I:t)
cx),
x2 )
~ P(I )
G a no+CD
lim G
n
cx)"
n)
n:d.,2 p 3,uuo
From lemma 3
p(G) P() 0
b/ Follows in a simUar tashion by defining Gn
3/
= (-
CD" n)
lim F(x)
= F(a),
JC.7a
x>a
tn~O,
sn> 0
If we define ~ (-
lim Hn H.
n?CD
IX)
a + t n)
interval (-
n:l,2,,3,uuo
ex),
a)
lemma 3
.6.
Problem 3/ Show that
F(a) .. lif ~(f) + P {Cal}
Where (a
is a.
x<a
aD
a)
Where F(x .. 0) is the
limit from the left #
Theorem 2/ To any point funotion F{x) satisfying properties 1,,2, and 3 of theorem
1, there oorresponds a probability measure peS) defined for all Borel
Sets suoh that for any interval (- ex>, x)
p(- co ~ x) F(x) ,
3/ A distribution
see Cramer p.
53
referring to p. 22 II
oontinuitieso
:> n
1#
P(Sn' ">
(n) ~ 1
whioh is a oontradiotion,
ex>
ex>
co
where
Notationg
[ ]
square braokets -- means the end point is not inoluded in the interval -ioe cr , an open interval"
round brackets -- mean the end points are inoluded in the interval, ioe."
a olosed interval.
to
b, inoluding a
but not b.
function
7
The interval is the set of points in
1\
i .. 1,
Theorem
4:
F{~ .. x ' ~
&:1
2~ ~
0,
10
2.
.3.
3:1
0,
xk )
II:
..
..
see p. 79 Cramer
0'
00,
0,
~).
0 0; ",
('1.'
x2 '
0 0' ~).
&2 + h2~
-----~-------6---~(~ +
h.t,
4>
h 2)
In R
2
P(I) .. F(~ + ~,
~ F(~, a 2)
-8-
,,-P(Sl" ~
-F(sl +
-I-
h:t,
+F(~.; a 2.9
1-
b:Ls
~i &.3 + h.3)
8,2 .. h2 , 8.,3)
3 + ~) + F(~D
8.
8.
...F(al , $2" a )
3
CII
~ F{~I
B- 2D
a,)
TIle proof of theorem 4 is by analogy with the linear case (theorem 1),
F(xl' +
Q)"
~ + 00) p(
CD" xl)
8:
.!J
~,
Exercise 7~
F(x, y) .. 0
..
if
x ~0
ex + y)
for
or
'1 ~ 0
O<x
~l
O<y
~l
- 9 -
But in fact
ai
F(O, 0) .. 0
F(O, y) 0 for all y
Fl (0)
F2 (7)
Proble!'l ~:
lit
F (+
00;
y)
o
.. 0
~l
elsewhere
.10 Delo
9:
Random Variable
1.
la
(.!ll
.!2J
()
1).
Cramer pe 1,2-4
axioms 1 and 2
--
Conditional Distribution
( .~ Y)
- -
~2'
It'
I xes).
P( Y C Tl J xes ) + P( yeT 2
,3-
p(X cS)
P(X <;.S)
If p(Y C T)
P(x
Co
1 X c: 5
CI
P(l CS)
SlY C T)
PCI c. SA yeT)
-P(YeT;
) P{ Y C Tl + T2 1 xes )
Pr(A
B)
~tA)B)
or
CD
>;
00,
y) then:
notation:
Def. 11 -- extension:
In the case of n random variables,
Note':
X $
2
..
I)
Note:
1.J.. ~
~,
X $
2
.,
The fact that the density functions factor does not necessa.rily mean
independence since dependence may be brought in through the limitso
y
eog. X and Yare distributed uniformly on OAB
lex, y) 2
O~x~l
O~y~x
- 12 -
Var1ables~
Functions of Random
Sf
xeS' ~ 8(:)88
is also a Borel set o
U LLU,--St
Sf
Sf
.=
Pr(yc S) = Pr(x cS i)
p(S')
S) p [g-1(5)J
Transforrnations:
We have X and Y whioh are random variables with distribution function F(x" y) and
a density function I(x, y)q
dX
'Ox
d~
'OY
~Y
1:1
a.;.
o~
X ..
III
'f1 (.;."
'f2 (..<,
~)
~)
2(a, b)]
IJ
is
.. J3,-
Solution:
Consider alsoW
III
X .. Y
z X+
(0, 1)
f-------44!~$ 1)
t(x"Y) 1
V-x-y
o Z+W
C02
Ja
X Z -2 W III Y
III
(:..,,0)
22
1
l'lI
...
2'-2
Density of
IJI
-1
Z X+ Y
WeaX..,Y
III
for Z
II
Q
III
Z ~
f(z"W) ..
Since we started with only Z, and "a.rtificially" added toT to get a solution" we must
now get the marginal distribution of Z (this being what we desire)o
z
z
F(z)
) F(t)dt
- 14
If X
'> 1, then
(z - 1) - 1
F(z)
to 1 - (z - 1)
(2-z) dz
2-z 2 .
2 . ~!L.
1
1
(2-Zi. +2'.. T2
~
z2
F(z) =T
1:1
f(z)
1 ..
o
(2_z)2
2
1:1
~z ~
(2-!l:.
-~
1 ~
.. z
2-z
1~z~2
~z ~l
See p.
24, ~ 6
in Cramer
F(z)
density
DoFo
2'
z
1
Joint density of
Z, W
fez, w}
2
1
1:1-
1
Oiz~l
... z-s.w'S.z
1
-"2
z - 2
~w ~
2 .. z
ioeo" consider in
R:J.
2
Y .. X
p. 167
X~ 0
Riemann-Stie~jes
Integral:
0.,
Let F(x) be a d.f with at most a finite number of discontinuities in (a,9b) and
let g(x) be a continuous function, then we can define
g(x) dF(x)
Cramer
PCll 71-74
.s follows:
Divide (a , b) into
~l <.
-s;
but as
0'
1i.
of length ~ !J.
S is increasing, S is deoreasing
b(
+00
Also define
) g(x) dF(x)
a-7 -00 a
b-; +00
"'00
b(
)
rf..x )
If F(x) has density f(x) which is continuous except at a finite number of points,
,
F(x ) - F(x _l ) -.t(JI') (Xi - x _l )
x i _l < x
xi
i
i
i
t~n
<'
-!', (x ) ~ (x)
j' .. ~
[int g(x)]
~li
g(x) dF(x)
lim
l
D
f (x i) ]
fjx
g(x) .!(x) dx
JC:l.$ x 2,
.,
nand elsewhere
is oonstant
If g(x) is oontinuous then this limit (the R-S Integral) exists o Also, if g(x)
has at most a finite number of discontinuities and so does F(x) and they don1t
ooincide, then the R..S integral existso
Def. 12:
=oJ.
~I
-I--..-------"~~ -
--'
...--- '
.----,r--__
g(x)
for such a discrete random variable, the R-S integral reduoes to a sum:
,J
b
.. 17 r
rr
is an intermediate
aa
t(X~ t) [ xi
x _
i l
.i X _
i 1
~ %~ r ~ Xi
g(x) dF(x)
.
J
g(~1 ~~6j ~) d . ~
x1 ,
HtJ,
For a del
II
Xt<:
F(x11 , x k )
of '\ see po B
It
.,
jrg('i'
a
0'
~) ~o.o ~ F('i'
~l ~
0
g('i"
~) f('i'
~)
0,
So
In
l1.
0,
.,
~) dxl' ~
ak
-00
ji
b
a.
+CD If we let b ~ .,. 00,
&.-7- 00
dF(x) 1
+00
d'i
~F('i'
0,
~)
-00
)
-CD
lim
dFx (x1Si x 2 )
a1-CD
b -7+CD
that we have:
xk )
.. 19 ..
a.., ..co
b-7 +00
R.k
+00
+00
0' ~) ~2'
0'
..m
- 00
Problem
f(>i' x 2, ' ,
6~
if Xl'
show that
U$
-2
ln Xi
Snedecor ch~ 9
Fisher ...- about pG100
Anderson + BancrOft -- last chapter of section 1
From problem $b
Z .. -2 lnX Y
-2 ln X + ... 2 In Y
integI'als~
Sokolnikoft -
Advanced Calculus -
ch.
.. 20 ..
Chapter II
Properties ot Univariate Distributionsl Characteristic Functions
Standard Distributionsl
A"
B0
[X ... a]
x.(a
x~a
Uniform (contimous)
F(x) = 0
x <0
F(x) ... x
1
rex)
C.
F(x) ... 0
F(x) 1
~x ~
>1
Binomial (discrete)
Pr
[x
sa
k]
.(:)pk(l
~ p)n-k
o ~p
1n
kaO
~ ( : )pk
(1 _ p)n-k
;; 1
is an identity in PI n
k=0
Do
Poisson (disorete)
'\
ct
00
e- A
'\ k
(X)
~ e->.. ~ .. e ->.. ~
kaO
E.
k~
k:
k~O k~
=1,
2,
.o~, 00
A >0
->..>..
.. e
e 1
[x . . kJ
r
. . ( r+k-l)
r-l
p
(1.. p)
k ...
l~
2, ..
oe,
00
O<::p<l
r is an integer
Example.
Draw trom an urn with proportion p of red ballS, with replacement, until
we get r reds out e The random variable in this situation is the number of
non-reds drawn in the processJ to have k black balls means that in the
F.
rex)
==
J'2ia
cl-0:>
Problem 71
<:
X <0:>
Prove that
t:~l
pr (1 - p)k 1
i"e., is an identity in r, p
Hints
Def. 14:
..
g(x) dF(x)
g(x) dF(x)
E [g(X) ]
Cl
-co
if X is disorete
Problem 8,
Given
F(x)
==
E [g(X)
1 +
'2'
rex)
dx
-OD
v=O
g(xv ) Pv
x <.:0
1/2
Q
g(x)
)(
ffn
S .
o
x == 0
t 2/2
dt
>0
.. 22 Find,
Def .. 153
E(X)
exists it is defined to be the kth central moment and
If E [X - E(X)]k
"'k0
is denoted
If E(X)k exists it is defined as the kth moment about the origin" and is
denoted
Exercise 9g
53
Theorem
"\0
Proofs
Let F(x, y) be the joint d.f Cl of X, Y and F and F2 the marginal dof
l
E [g(X) + hey)]
III
Theorem 61
II
Proof.
Corollary.
E [g(X) hey)] ..
E[g(X~ E~(y)J
'\
E(X) .. mean
"'-2 .. E(X ... lJ.) 2 .. variance
=:
~ .. E(X _ ~)3
""4 ..
E(X ..,
~)4,
etc.
for the normal distribution -- N(O, 1)
..x2/2
t(X! 0, 1)"
.J 2n
_1
{2i"
00
_..k
E(X--)
1
.---..-
=0
(>
by a
"s~etry"
argument,
23
E(X2)
E(X 4)
III
dx = 1
{2n
3
or in general
E(X2n )
2
2n -x /2
III
{2n
dx = 1 :3
5 ..
z: (2n - 1)
~" ~,
u.,
ilk
CD
:>O~ ~ -
Proof~
Example I
N(O, 1)
~k+l
=0
co
k=O
co
III
~ _~(_r2"",,)_k_ _
k=O 2k- 1 (k-l)! .2k
2
{r )k
~ ~(:;)k
k.
_r 2/2
III
::I
==
exponential series
"
f(x)
1lI..!.-
V2it
-x /2
Problem 91
Find the moments of the uniform distribution and show that this is the
only d.f. with these moments.
Theorem 8a
(Tchebycheff's)
Pr [g(X)
~KJ ~
E g(X)]
K
00
Prooh E [g(X)]
g(x) dF(x)
't KJ ? !.(X)
S g
dF(x)
.f
dF(x)
III
'.- Pr
Corollary II
[g(X)~ K]
E G(X) ]
K
The above (th. 8) converts readily into the more familiar form
Proof,
g(X)
~(X)~KJ !
~K
III
(X .,
Pr [ (X - lJ.)
(k 0)
J~ TI2
'" k
- 25 -
X
n
II
O'2x = np(l-p)
Proof &
[I :.r - pi ~ k
Choose s
=5
0"]
(,
or k
kJ p(l-p)
n
;l
1
II -
-J8
= e or
n = p(l-~l
8 e
Hence 1 n is chosen this large~ from the corollary to theorem 8 the stated
probability inequality follows o
Note I
n=~
8 s
Characteristic Functions 3
Def o
168
Characteristic functions
X(t)
=]
eixt dF(x)
or since e ixt = cos xt + i sin xt
CD
(X)
X(t) =
-00
cos xt dF(x) + i
- 26 -
(t)
dt
J ix e uro dF(x)
= v (t) =
m
-(I)
CD
' (0)
= i
x e O dF(x) i ~
-00
in general
the moment generating function operates in the same manner, except it does not include
the factor tti" $ and is therefore not as general in application,
MaF
= M(t) = E
(e R )
00
= .~
ext dF(x)
-00
also true.
Examples I
1.
Trivial distribution,
Pr
[X = a]
(t)
CD(
III
-00
20
Binomial.
1
xt
eJ.
The converse ~
- 27 co
3.
(t)
Poissona
k
eitk a-A _A
'"
~
III
kl
k=O
411
Normal (O~ 1) S
(t)
1
III -
.J2n -00
1
=-
{2n
Fn
.~... e-
2
x
2' 2itx
dx
e-
dx
setting y = x - it*
(Note the term in curly brackets is the integral of a normal density and equals one.)
If X is N(O, 1)
(t)
III
e-t
/2
2t _t 2/2
= - reI'
.)1
.),
(t)
= -e
(0)
=-
_t 2/2
2 _t 2/2
+ te"
22
E (X)
=i "
=1
.. 28 ..
Problem 10 s Prove
Problem 11.1
Show that
iBt
X (At)
Proof:
Pyet)
I:
eitjJ. x(a t)
= e
Def 17 i
<)
i)
it
jJ. e
;:;e 1tjJ. - .
K( t) :Ln ~(t)
T-
i)
- 29 Note I
Note,
Theorem 11:
If
J1.P
X2J
0 UJl
Xn are independent
ran~om
variables, then
tr~
.~x.-""
.i=l'
. J.
Ai
=e
n
n
therefore we could say that Y is N(] j.(.i' ]
1
ai)
To justify this last step we need to show the converse of F(X),"""? X (t) ioe"
X(t)
~F(X)o
sin ht dt
- 1 b<.O
III
t
Proo.t'~
that
J(..(, ~)
~f3
III
a:>.
III
(-
cJ
0
+1
e-...(U
e-..al
cos
h '=0
b>O
~u
du
sm r U
u
du
'.,('>0
/
Note I
- 30 ..
J.
III
J.~+'P2
f~ dp =2~2
Let
djl
= arc
then J(.<., 0)
{3""'70
tan
~ +c:
e'<'u ~ 0 du ... 0
arc tan 0 + C = 0
o + C ... 0
h>O
-T
n
... "' 2
Theorem 12&
h<O
It F(x) is contimous at
1rfil
depending on h + or -
00
\-7 00n
Note~
1:..
lim
a - h, a + h, then
i.t
-1
....ita
SJ.ntht e
(t) dt
.1 .
-itxIt (t) dt
itx
(I)
lex) dx
~el~
16-1
Combining this with the above theorem means that given (t) or lex)
we can determine the other.
Prooll
Define
1
J=n
;T
sin ht e
-ita
wet) dt
-T
T
1
=n-T{
sin ht
-ita
Ell
CJ
/tx d
F(X)] dt
or)
.l.
n
...ita
tee
...
J[J
The
T( sin ht
-00
1
a_
No1>e!
'-
s~
ht
S~ sin term is an
;it(x ..
Notes
_~_(
J
,r.b
dt
a)
d F(x)
i sin t(x..a)] dt
_~
itx
dF(x)
0. Ja2j
a}
1 fj
~ sin t(x-a.hl
sin t(x-a+h)
dt d F(x)
---11---------
both
= ... 2'n
x-a-h<..O
:1st part
n
'2"
I.. .
Xi_-_a_-....;h>O
both
lit
I
J 2nd
I
part""
I
I
1t
rf
T
sin t(x
~a
~x~a
+ h) dt
+ h
T( 1
) t sin t(x - a .. h) dt ..
o
.. 0
elsewhere
a+h
1
Ja-
There
alh
a-h
CD
o.
dF(x) + ..
1l
dF(x) +
a~
0'1 dF(x)
.. F(a + h) .. F(a - h)
Proof of the second statement in the theorems
F(a +
h) F(a h) .
2h
ex>
sint&
ht e -ita rI.(t)dt
'P
-00
= 2l n
tea)
<X>J lim
_
h-.,o
sin ht e...
ht
ita
(t) dt
"-~=1
therefore
f(a) ...
-J
'2i'
-ita
e
(t) dt
Problem 12iLet Xi (i .. 1, 2,
eo,
a-l
f(x) ... a x
a>O
n
-rr
i=1
X
i
B(X[r]) = E [X(X-l)
Define
F (t)...
.1
ex>
(x-rotl)]
- 33Find
F*(t) for the binomia1 and use this to get the factorial moments.
-It I
Omitted -
F2 F
3
(11. 2 3
ouu
fl
we go from the Fi to the i - .. observe that the i tend to a limit ~or exampl,
the normal approximation to the binomia~ -- observe that the limit is itself
a charaoteristio funotion [of the normaJJ .... then go to F by previously
disoussed teohniques.
Theorem
l4~
If'
X:t."
moments
(say mean
IoL
vn- (*
~ Xi
IJ.)
tends" as n
normal distribution with mean 0 and variance (]2
1 is asymptotically N(O, i)
(t)
n
rn
~ (X...
~
IJ.)
~ (Xi""'
rn-
IJ.)
~oo"
to the
- 34 ...
it' we expand
(t)
where
(t)
i2~ t 2/2
&:
1 + i ." t +
&:
~ ~O
." = loLl
+ remainder
= I0Io2+
as t?O
In
&:
rn lJ, t
:: i IoL
Note: In (1 + x)
rn
x2
= Xae T
+ n In
(i>]
$6(*)
t + n In [ 1 +
ifi..
al :::". ) ]
lrn
2n
x 3 x4
'1"
.. "4
x
= x -"7
r6
+ In
+ R
.~......,
where
setting
ln
0{
~n
X!'I
vn
.rn
(t) = ..
1.
i/
i IJ. -
as x ~O
2
2
2 t
t
.. (IJ. + a ) ~ + R (~) we get
t:n
'In
2
2
2
t
~t + n i lJ. .., (~ + (] ) 2n + R (-)
[trn
(i j.L.i)
rn
3
~
It]
An + R ~t)
+nR(~ +V~~
n~CD
In
r (t) =- a 2
n
it2
"
lim
n-too
rA
'Yn
(t)
&:
e-
-r
.,.
Y2
denote this by
2 t2
lim
nJ/to.
n
~
Tn
Rill
since lim
n-)ro
\-;:::F/"" = 0
./ U
lim
R~ t
n~
Q)
~) l
&:
t
0
since as n~, ~~
which is the characteristic function of the normal
distribution N(O, (]2)
Note 8
See Cramer p
214""l215.
tI
The theorem says, for any given s there is some sufficiently large nee, a, b) such
that
b( -t2/202
a.
btl
j
dt <::::s
Problem 15 ..- says that the standardiZed Poisson random variable is asymptotically
normally distributed N(O, 1)
. ..,
.,
Cauchy Distribution - see problem
differentiable at the origin
lim
b~CD
- \t
is not
.!.
n
a"-7CD
=oo-co
therefore E(x) does not exist for a Cauchy random variable, even though the distribution is symmetric about the origin"
Theorem 15~
(Liapounoff)
If ~I X2,
2
0i and with
e&o
P~ = E
I Xi - i 1
8
~p3
(~
VI -']00
then
Y
"O'i
Proof l
an
""
~X. -~J,t..
:L
)
:L
=c
;r:L
is asymptotically N(O, 1)
1/2.
."
3/2
and variances
- 36-
III
k xk- l
III
...
III
k~t
then
Wet)
0 ~X~
&~
O<x~l
III
Xf2
III
III
x<O
Replacing (s) by (-t) gives the m(Jgof .. , or by (-it) gives the c.f 0
[eitx ]
a;
\Ie have previously noted that if X, Y are NID with means jJ.. and variances
221.
=:
normal
name
density
1
n/2 ..
2
r (~)
n.
;~
...x/2
.Cof.
(1 ..
a
n
2it)~
2n
- 37 ..
n
~....... Xn are
.- it'
NID(O, 1),
x2t
is y.,2
= X; + X~
aA
-ax
"'-1
x
rCA)
- = Pearson's
student's t
(n)
-_ t
~(~)
rm
r(~)
=XJ n
y
--
2 nil
X~
where
n-2
Ln ;;.2)
( n",>l)
X~
r{~)
F(m, n)
r<;) r()
i 1
n/2
( ~)
---
X m+n
(1+ !!! x) T
n
.,
2
2n (m + n ... 2)
n
n'"
m(n_2)2 (n-4)
~
X>O
-
if X, Y are independent, X is
Fisher's z is defined by F
= e 2z
2
Xm,
see Cramer p.
243~
Beta
13(p, q)
r (p+q) p...l
q-l
- - x
(l-x)
rp,rg-,
__ putting P
= ~o
= ~,
we get 13
.~.
p+q
~ F
= _n_ _
l+!!F
n
Jil!!'YTr1
38 Gamma function,
rep)
rep)
III
n-l
x
(p-l) r(p-l)
i f p is a positive integer
r(p+l)
lim r~h)
p
n>O
dx
III
rep)
= (p-l) !
Stirling IS Formula
=1
h Fixed
rep)
P~
Problem 178
2.
III
Problem 18g
Cramer po 319
No. 10.
(j
).
-39Chapter III
Convergence
Convergence of Distributions:
a. Central Limit Theorem (theorem 14)
b. Poisson distribution - (problem 15) -
(a, b)
~>...,y.
K!
= >..+a{X e
LetK= >"+VAx
-rr
X+6x aK + l - h
iA
ax
b
..
(a,b) . x=a
.-A AA +
xl?:.
111-
rr
(h + x~)!
nl
where e{n)--7 0
as
n --7 co
III
1 + Q(n)
-402
Notes: limit
= e-x
X 700 (1+
limit ln
lfi3:1t
Xvr -
:::J
{i
where 9 (X) -7 0
as 1..700
r
b
Note:
.l..
f2;a
2n
e"'x
h2~
e..
/2
dx
For a similar proof for the binomial oase, see J. Neymam First
. Course in statistics, Chapter 4.
:_n
~
V 2n
is A ~1(0,1) as
n-7
00
X-o
J~~
is A NCO,l) as n 700
X has mean...P....
p+q
Proof:
Omitted
1
1 + qJp
-41Problem 12:
a.
b.
Problem 20:
Define E
[seX) (Y] ..
where f(x y)
g(x) f(xly)dx
-co
::I
f'(XJY>!f'l(y),
DEE [g(x>Iy]
Y X
Use this to find the unconditional distribution of' X if
Y is Poisson (A.).
Convergence in Probabilit;y
Def. 18:
-p
-t tt
indicates converging
in probability.
J!h
p> c i f given
t,
8 there exists
>N
Pr[\~
cA>~e]< ~
p-1
Ii
Theorem 16:
If' ~,X2' ... 'J!h are a sequence of independent random variables with
means l-Li and variances
n
cri then i f
Lcr~
1 1
---""2- ~O
n
then
Yn
~ ~ (~ ~ - j
-42Proof:
Pr [I
given
~, ~
choose
"
Now
oi:'n
Put K
IOl
'.
iii
v'"f
V2
{6
n
)'
~ ~< ~
iii
o~ ~
-7
Gl)
(~ f' o~
n
)1/2 <.. s
Pr[1
as n
~ ~
(Xi -
'1.~ ? ~ J< ~
ai
2
o
=n-
~ IJ..
xCt)
In
IT t
E(e ~"')
i(~ IXi)t
l(t)
E(e
. (t)
==
n In (~)
where
~ -7 0
Lx(i)] n
rL
= n In 1 + i IJ.! + R]
n
as t
-1
.v. with
-43-
In
.
(t) .. -1IJ.t + n In(l + i lJ.! + R)
1-1.l.
n
Note:
In(l+x)
where
-:UJ.t .. n ~i
III
R~X) -7
(t) -) 1
as x
-j 0
RU(!)
...ntn.--to
so
henoe
Rn(1)
= in t whioh as n-f (I)
x + R(x)
as n
-t 00
!-lJ.
but i f (t) = 1
F{X) = 0
x <.0
x~o
III
Questiom
X:J.'
X2, X ,
3
Do
II
lJ.
~3
Do
2
2
~ 2
If
1>&0'
Xn
." .
lJ.n ~lJ.
GO"
,in ---762
i.e ., does ~7 I
with probability
-1
n
1
1- ....
-7
-44,"~ X
E \'" X
l n
-j 0
:I
0(1 - 1) + n2 (1)
n
n
as n -)00
:I
r.~1-t 00
Problem 21:
Problem 22:
r=
Theorem 18:
u"x.a
Let X.,,~,
be a sequence of random variables with distributi~
functions Fl(x)" F2(x), "0' Fn(x) ~F(x). Let 11" 1 2, ""Yn be
a sequence of random variables tending in probability to c.
Define: Un
:I
~ + In;
Vn ~~;
and Wn
:I
~/Yn
see Oramer p.
:I
/1n
-cl<::.-
xn + Yn
.:::.
Xj
In .. c \ 7
:I
..
..45-
In ~:
-1
F@x ....
0 ..
e -Soyn::- 0 +
0 .. 6) ....
P(~ + ~ L. x)
~ p[ V
&,
thus
.
f<F{X" c .. e)
= P(~ = x ..
~c - ~)] = Fn2 [
:=I
[xn ~ x .. (c + el J :::-
Yn)
x - c ..
~.o(x -
c + s) -
where n is chosen so that when n;>n Fn(x) .... F(x)(i in the vicinity of c.
2
2
Therefore" in Sl
0 ...
e) ""
6) -
NO~ing
.... 8 ~ ....
i?hat ,Pr[Un ~ x
i .. i ~ F( x ..
p( 8 ) +p( 8 ) ..
~ F(x -
+ e) +
0 -
6}<
6) ..
~(x
c)
i + 0 - F( x ., c)
= Pr [Xn ~ .xl
i +- i . . F(x -
0)
~i
:"" F( x ...
~
c + e)
F{x .., c)
II(
F(x
0 .... 6)
F{x .... 0)
III
~ ~-
Pr
[xn~xJ
c)J~~
L.
.. F{x ..
~ ~ 6
i
i
o)~- ~
for n7max(n l ,n 2 )
...
.,max
(~"n2)e
we can write:
-46-
:'heorem 19: If X
n
g(X )
n
ProBlem 23i
>0
0,
P-} g( 0)
XJ.,~,
000
sn =
whioh is in the Wn form
~(\ _ 1)2
{n _ 1)(]2
::I
n
n -1
=-
~, val' (]
-47-
p ) (]2
i.el)'
) 1
(x - J.L)2
2
(]
':j
J.L
(I ... t.l.)2 _
Therefore.
,
-7
J.
:;J
pI
as n -7
ex>
2
and sn
by Khintchine t s
2' ~
p
(]
(]
s
and thus by another application of Theorem 19 -;
Note:
p ., 1
Re Theorem 19, see: Mann and Wald; Annals of Math. Stat., 1943;
itOn stochastic Order Relationships" -- for extending ordinary
limit properties to probability limit properties.
Mise .. remarks#
GO
a) ff La +
Q (x -
a)]
~ 9 ':=-1
=:
10
where R
= (x -
-48R
IS
ft
then i f
[a + G{x ..
is continuous" as x --7 a
f~
lim
III
x...-.)a x-a
If there exists an A such that fA g(x) dF (x)
Remark:
0)
..
for n
)
g(X)dFn(X) - - ;
...co
Question,
r::;
and ,
<s
g{x)dFn(x)
Ref .. Cramer, P4 74
g(;)dF(x)
...( 1 ) '
Counter-exampJ,elt ~
1
2
;-; e- x /2
Define Pn:=...L
~ 1-n
In is normal (0,1) except on the interval. (1 - ~ , 1)
and
III
X:L'
.l0-
IJ.)
= ,.,.
~
is asymptotically
(X
X".)
X~l) as ".-7
CD
X2,
"0'
III
<s
A sequence of r.v.,
X:t'
--t X a.e.
X2, ...
i f given
8,
6 there
p ) c
with probability 1 .. ~
-,=:
1 with probability
n
the XI s are independent.
(1)
To show Xn
p -j 0
.. !n
(2)
~-/-10 ae ..
P.r
D~ "'"
0 )
<
00
III
IT Pr[~L.E]
nmN
Note:
l - x <:::.. e-x
= N,
N + 1, N + 2, ...
CD
= IT
(1 -
N;j)
j=O
0
~x
=1
-50therefore
[I Xn - 0 I < e
Pr
therei'ore ~
= N,
N+l, N+2, ]
=0
0 a.e.
Problem 25:
If
then
=0 with probability
=1
--j 0
with probability
1 . . ,1~
1
7'
a.e.
Problem 26:
= cos ta
x(t)
IJ.,
1/
and variances
vi ' a~.
Find and
Kolmogorov ine9Wit,V
Let
2
and variances O'i"
IJ.i.
Then
Pr
I~
n n
[~~
be a sequence- of r
oVa
Xi
~ fl \ . q
(tai
)1/2
Exampls3
J1.
X2
Pr
[J
Xi
\-<
\ JS. + X2 \
<
Y2K ] '/
1 -
with means
Theorera 208
oo.
2
and variances ai' then, i f. L\
l-7
l1n - 'iin
2/,2 converges
ail.
0 a
where jj;
= 1n I
J.!!! !
large numbers
l1i
Let A. be the event that for some n in the interval 2 j ...1 4.n '" 2j
J
-52-
Pr
[J
t<l"- ~)
17
nJ
~ ~ L<x" - jJ.nl
I I n - ~JJ
e
'7---(I <7~
I
j 1
2 -
Pr
j l
2 ...
{X
L.
PI'
)1/2
<7i )1/2
x..
-J.'"
h1
L.. If-
<71
IL(~ ~ ~t
(JS. + X2 ) .. (!1 + ~2
<::... If;
t:,;
~
(2 + 2)1/2
<71 . 0'2
..:..-_--..;. <:
{~ <7i)1/2
upp~
~ ~: 2~j
L.. 4.
-2
1=1
2j >i
Note that
1-':'2"""
2
2j 7i
16
38
III
henoe
I l
Pr
'X --1
Proof:
fJ.
a.e.
is immediate sinoe
\' O"i
L ":"2"
,,2
III
I!2
Aj ]
X
n
1.i om.
III
III
i f lim E [X
n-ja>
- X
J 2 ---7 0
Cramer -
~ a>
-542_
3.
Note:
Ln
~n
=1
e.g. Toss a ooin until heads comes up .- oount the total number of
tosses required ( II! n) .... bank pays 2n to the player.,
E(x)
III
for a fair game, the ' entry fee should be equal to the expeoted
gain, therefore, this game presents a problem in the
determinat~
~ ~n~:in ~ l>~}~
1
tha.t is, the game 'lbecomes .fair It if the.. entry' fee 'is .n. 1n n.
ESTIMATION (Point):
Ref:
!,
8 ~
9 s
n -- parameter space
Example~ 1;
if Xl' X ,
~ ., Xn are NID(~,
i)
then
2: F(x) Fs
7'
1
co
x dF(x)
"'00
from R to
k
-f.L which
<: co
g(Q)
d(!>, Q] with
W .. 0
Poet;)
20~
~)
... E
W[ d(!),
]1 . .
co
S
-00
W [ d(!),
dF(~ ~>
- 56 -
d~(X)
... f
R(d*,~)
E(X _ ~)2
101
~
n
A t'best" estimate might be defined as one which minimizes R(d, Q) with respect to
d uniformly in
Q.
..
g(Q)
o
Hence amiformly (supposedly) minimum risk estimate can be found only if there
exists a d(x) such that R [d(x), Q
0
J..
stopped clock which is right twice a day, or one that runs five minutes slowo
Since a uniformly best estimate is Virtually impossible to find, we want to
consider alternative
WAYS to formulate the problem of obtaining best estimates~.
I.
J...
g(Q)
t de!) -
g(Q)
J2
invariant estimates
Let heX) be the transformation of a real line into itself which induces
a transformation h on parameterspaceo If d [heX}] "" li[d(X)] then
d(X) is invariant under this transformation.
Example~
< 00
Ii [E(X)]
An estimate d(X) of
~ is invariant if d [h(X) ]
d(aX + b)
Q.
a E(X) + b
101
Fi [d(Xi}
a d(X} + b
- 57 -
Dei'. 22: Estimates of g(Q) which are unbiased, linear in the X. and which among
- - such estimates have minimum variance are b .l.u.e.
J.
Problem
2~:
Xl' X2, 0 " OJ Xn are independent random variables with mean IJ. and
and variance (12
I is the
show that
Problem 29~
Xl' X2,
W [d(X),
0,
QJ
bolou~e. of IJ..
co
b [d(I) - IJ. ]
III
a"
d(X)"
(note:
X find R(X,
b.
d(I)"
X+a
[note~
(1)
the answer depends on the loss function constants only,
not IJ.)
j..L,
-co
Comment on this
problem~
An orthogonal transformation
.- is a rotation or reflection
-- l. .., ~ where A is orthogonal
.... I J 1= 1
.~
...-....wyi = -"ix i
...... For
II
+ a.l_x :
~ a~ ... 1; ~a'jakj
...
j J.
.Lunj=lJ.J
R(~,
Q) .. E[W(d, Q)J
f,.. f
/
V [T(X), Q ] dF(!)
et
0'
In)
y ... T(x)
y ...
Then R [d(T). 9}
f'l
[T(X). 9 ]
~(Yl!
000.
Y</
~
g
II.
r'l
S
III
T(x)
of Y2' Q"11 Yn
given Y1 1
[Y1 , 9 ]elF(Yl)
Defo 23:
If
o.r
has a known "a priori" distribution H(Q) then the Bayes estimate
g(G) is that d(G) which minimizes
ft
Example:
x ... O~
( : ) pX(l _ p)n-x
SR(d, Q) dp
- 59 -
=~
2
[d (x) - 2pd(x) +
x=o
i]
xJ(:~xli pX(l,
Note:
(a{)b
) p 1 - P dp
- p)n-x dp
a b~
(a+b+lj1
o
Using this evaluation on each part separately we have
x 0
[2
nl
xl (n-x) 1
2d{x)nJ
{x+l)J(n-x)J
d (x)xJ(n-x)J (n+lp - (x)J(n-x}l - (n+2H
+
1:1
~
a
~ r
. 1
2{) _
n+l ~ 0 ~d x
1
n+l " 0
x
nl
xHn-x}J
2
[d (X)
2d x) x+l
0
n+l - n+l n+2
~(x){x+l)
(n+2)
nrr: '
X+11
d(x) ... n+2
(x+2)J(n-x)J ]
-(n+3JJ
(x+l)(x+2)
]
m+l){n+2Hn+3)
{~+1)~X+2~ _(~
'\ 2]
n+2 J
+ (!:!:!)2 +
n+2
\n+2) n+3
x+l {n+l...x
}]
+ n+2 \.Tn+2)(n+3)
This is certainly minimized with respect to d(x) if each term in the first summation
is zero -- ioe 0 if
d(X)
.
Problem 30:
if d(X) ..
1:1
X;l
n+2
io R(d~
average R...
p) dp
if P is uniformly distributed on
R(d,? p)
os
E [d(X) ... pJ 2
(o~
1)
- 60 2.
Minimax Estimate
Def~
24:
(9)
i;Je., we get the min (with respect to d) of the max (with respect
to e) of R(d, 9)
-- or we take the inf (d) sup (e) of R(d, 9)
dl(x) is minimax estimate since it
l---"L------~
_ _- -
R('1. 1 Q}
'--------
.au!~ Q)
25:
Problem 31:
II
Der.
~-
d (X) is consistent
n -
it:
d (X)..
) 9
n p
and J(d
HO
or that d~drl70)
atdnJ -
2.
lim
n-?co
if
--
<:. 1
.61METHODS OF ESTIMATION
-.
A ......
~
~thods
ofdi>ments
(Ko Pear$on)
sample moments to
Xl" X2 '
then
3'
.It.
'$
population
note: This method yields poor estimates in many cases -- has very few optimum
._ .properties
.
B -- M9thod of Least Squares (Gauss -- M:l.rkov)
let X ... {Xl' X2~
all ....
A ...
<I
a nl
0"
i~e.~
s~n
A is of rank
a ns
s
..
j .. 1
aijQ j
also
Def. 28:
i ..
l~
2,
<I
if
OJ
III
* minimizes
00
e -
a is 98 )2
(Gauss -- Markov)
With the given conditions on Xl' X2' Q$ Xn the least squares estimate
~* is a best linear unbiased estimate {belouve.} of ~ ~
- 62
Proof Theorem 21:
ref:
it' we write
g* ..
= C.1
c1
= (!. _ AC-1
A'
1
AC.. A'
1949, po 458
BiometrikaJ
At X
A!. + l
with respect to y,
= (! ..
Placket:
Lehman
where
1:1
AA
=C
since it ia
symetrio
i.
A ~) t (!.
(! ..
AC
A ~)
! .. Al)' (! ..
!.)' (! -
1
(1" AC.. A'o!)' Al
AC-1 A'!) + (Al) , (Al)
'!r -
1 A
(! ..
- (Az) t
-! Az + !
eClg4
since
'-1 ' ,
) A Al
A(C
(Cool)
;:I
1:1
..
Cool
Al. +
! Al. ..
= A'A
=I
happen i f
Az ..
or writing
-I)
(I
a. 9)2
0"
1S S
.!
formally minimizing
in the usual fashion by differentiating with respeot to
the 9
j
n
=
-2
~
a .. (Xi .. a
9 .. () .. .. a ..9 .. 0
a. 9 ) .. 0
'1r
11 1
d~j
i
1 1J
1J j
16 S
'dl
(t
;:I
1, 2,
j ..
0, a
~.
=1
a i .x.
J J
(~
i. 1
or A X (A A) Q
aijan
91 + '-'
(~
i 1
aij&is) 9 s
. 6.3 ...
To show that
so defined
(ioe.,
' ) is B.LoUoE e
'"' C-1 A!
B~m!
B AQ '"' Q
thus BA '"' I
note that o-2BB ' is the covariance matrix of B! -- the elements in the
diagonal are the variances of the estimates ~ ...., we thus wish to minimize these diagonal elE;1ments (by proper choice of B) subject to the
restriction that BA" I
BB 8
..
(0...1 ) I
..
III
0-1
0 '... 0 or (C.. l ) t
\ AOf A '"' 0
* are
example:
if
B~LoUoEo
JS.,
X2J
.~
common variance
ci
let A
.,
here
0...1= ~
'"' 1 ~ x... f
n 1=1
= 1"
2,
0,
col., ~
and assume
~ ti
1
using theorem 21
lOl
].
64Aiken extended this result in 19.34 to the case where the X. are correlated and
we mow the correlation matrix V (up to an arbitrary mul tlplier) -b.lou"eo are also least squares estimates which are obtained by minimizing
(! _ A f~) f
V I
(! _ A t~)
ref:
c.
Plackett:
Biometrika, 1949,
p&45~
Defe 29:
Likelihood function
L
III
f(JS., X2 "
I)
0,
III
p(Xl, X2 '
I)
."
L .-n-t(X., Q}
1.1'
or ln L
II:
= 0
i ..
1, 2,
Q ,
~ ln
1
(X., Q)
1
2.
The
dQ
we assume that
"lln (X
';)gk
4.
dk In t(Xi ,
09k
Q)
<
i , Q)
exists for
= 1,
2, .3
00,
00)
50 E
ta;nQfi~2
i.e., it is bounded
...-
O<k2< 00
Q) dxi k2
Theorem 22:
"'-Ma'
In L, has
1-
,In L
0
09
1\
2- 9 is consistent
3-
rn (~ . . 9o )
Proof:
1-
f
n
L(a
[o~l~i
since
In
and varianca
In f(X1'
has a maximum,
\1~ L.. 0
at this
if
max~
is consistent
summing each term on both sides of the equality, dividing by n, and doing
some substituting
1
-
aIn L
e- 9
1
...::i.
n i=l
s:: -
d In f i
~Q
~ [~.21n f i
+ - ...::i
Q=9
n i=l
o
Q2
9=9
( 9-9)
0
z
(Q-Q )2
+ - ~ F (x )
0
n i-l 3 i
2
!..n
a In L
~Q
written~
(9_9 )2
B + B (9 _ 9 ) +
0
BI)
....
20
-----~-----------------~~--------------~---
f(x i Q) dxi 1
..00
- 67 ..
-----~--------------~~----~----------------
Let S
= the
~O
B2
2
.. ..k
I B0
IL.. ,}
We can f'ind an
~~
Bl
n j given
Pr[S] >1 - s
8j
In S the right hand side (rohos.) of' the Taylor series expansion is to be
considered"
Consider: Q = Q + 6
o
~ B2 62
~ F} (1 + M) ,., ~ k 2 8
k2
So that" 1
Considering:
Q ... 9
<
the l"ah~Sc
2{1+M)
~ 0
2
8 +
6" 9 0 + 6) and
.. M5
f'orthe same
__
~ 0 "
- 8
'? . .
letting z 1
= -(M + 1)6
2 +
k 26
<~(l+M)
> (1 - 8
~ a"'al~
L .. 0 has a
ln L has a maximum
- 68.1\
Q
is then the maximum likelihood estimate and the solution of the equation
! ~lnQL
n
~..,
whioh yields
($
... B + B
l
0
Q )
0
+ !2 z B (Q
2
Bo
... _ _
co
_Q
)2 .. 0
_
.
B +2" B2 (Q - go)
1
k..r;;
We lmow that:
1
]
V [ Bo J ... ~ E
therefore
[nB o
is
[d d
ln
AoNo(O, 1)
Bl-~p->~ ...k2
9-Q---0
o
p
O~
2"
1
ln
III
E[
d~
f.i
----------------~----~-------~------------Example~
f(x)'" a a-ax
Find the moloe o of a, and its asymptotic distributiono
n .
L .. an a-a
Xi
.. 69 n
In L n In(a)....
xi
1
..........
..
i-l
n
.~~
therefore
= -1a ,
c>
E(x)
=-1a
'V
or a .. -
is A N(O~ 1).
or
--------------~-----~-----~----------~-----
Problem 346
-
2 ..a 2x
f()
x a e
x ,>0
and its as;ymptotic distribution.
Problem 35~
--
Xl is Poisson ~
find the
~lc:>eo ot
(i = 1.1' 2, "
.J
n)
- 10 Problem 36:
(0" a) ..
[not by differentiating]
_... Correct it for bias and find the asymptotic distribution of the
unbiased estimate (~!, e
-- Another unbiased estimate is
* .. 2 ..x
NoteZ
Remark~
If dn(X) is
afn)
A N(j.L"
and
F" [g(dn ) -
g(-IJ.).J
is A" N(O,
[g'(~)J2
.;, 0 at x=1J.
(12).
"---~)
rn (g(d~
(j
------------------------------------------Multiparameter case:
Theor~~~
"
"
and with
- 71
probability 1 - t, the
of the equations
0
III
and furthero
mol.e~
9 , Q2" ,
1
~o
~s
13
9s
."
V"" ... n
~, S
where
i 1, 2,
~o
"0
of
and variance-covariance
13
13
l
ri
In.1.\
Vlni
d
( dQsQl
s 2)
E
9 (0
(:lln f \
o
E (d2
;:
oa
d::l :
L
i ~ ~l~iL toa
(Note:
In f i
CI
i 1" 2, , s
From theorem
22:
lnL...
~
1
ln f(x i ,!)
2
. E(ad ~
Q
ln
_ E f:.d ln f i ) .. E [-( d f i .) (.
C-d Qj
Qk
~Qj
~dln fi)J
Qk
- 72 ..
Now we oan write:
------:lJo)
p
n
b ok
J
!.n ~
1
II!
The maximum likelihood equations can be written (ignoring the second degree terms
in the expansions):
~ ~
A ~ 9 )
.. B (g
1)
B -p....--~> E(BJ
For la~ge n~
[ J -1
B E(B)
-B
o
or:
p ) 1
(~ -
B [E(B)]-l [E(B)]
13
($ - gO\.
[E(B)
-~(B)
J -1 Bo
gO)
13
(E(B)](Q -
(P)
"
V and is non-singular.
CI
S' ..
Variance-covariance matrix of
gO =
E(B) -1 [ var-.cov
V
f Xj afj b
(
...
'fCE'
a
-ax b-l
x
X;>O
a 0
ln L .. n b ln a ~ n lnreb) .. a
Xi .., (b ... 1)
defining ~ ""
b'70
~ lnx
. n
i
- 73 ~ln
Lb."
....
n . . a a ." ..
a x 0
1.
dln L.
n~b
ln a -
r/i/f~t
+.~.o
Xr. 0
ln (' (b) )
defining
lnb-F2(b).lnX-~"
Note:
Thus we have:
.....
'd2 "In
da 2
~21n L
b
oaa
'?J2 In L
~b2
1
a
7
n
V -
IvI-
-1a
F (b)
2
[+ ~ F~ (b) -~ ]
note:
1\
of
of
a 1\
a,P
JVI
J1.!.
Iv/
2
F2'(b)
a
n [bF ~ (b) - 1
... 74
...;n
b);
(~
means
F~(b)
:l
b
a
i
b F (b) .. 1
2
Exercise:
f'(x;~,il (] ) ..
j2"";'-e(] >1
and
j.l.
(]2
(x - f,/:)2
2
20-
..., The
....,.
mol~e
of
a and
~ (xi" i)2
1
are x" - -n - - -
~2
~3 '6-2
.. oo.(x <00
Note~
j.l.
(based on n
independent observations).
38:
an unbiased estimate of
X is Poisson ~
A"
&
Y is Poisson ~ j.l.
Find the m~l<!le(l of
~? find a2
x
why not?
j.l.
.. 75 ..
Check that the same results are obtained by expanding
about
Xi
Problem 39: a
ti'
~
in a Taylor series
ClO
is
a">O
b70
j 1,2, ... ,n
matrix.,
- 76 ..
D.
UNBIASED ESl'IMATIONe
Theorem 24:
I:
for all n
where
)~ngf
JS.,
~,
ou.t
f(X.? e) 0
2 '"
t~
I:
'3.
40
d(x) f(x" e) dx can be differentiated under the integral sign with respect
to 9&
5. 0<k2~
Proof!
CD
~quation,
we get
- 77 -
wh 1.0ch ca n al so b e wrl.Ott en E
k(!) 4dl~
L
(E[S(~)
-
Note~
O"d(~)
d<l) ] )2
2
O"s (:1)
(S<!>l
E [d<!:)]
I:
1 + br
Oln f(X)
dQ-
and d(X)
] =0
or
now
therefore
[1 + bR(e)] 2
'D in f (1)]2
n E -'09
Since r = 1 if and only if the random variables are linearly related it follows
that the equality in this result holds if and only if d(~) is a linear f!J.nction of
S(.!)
!2rC'Jrlplea
JS.,
X2 > ~I);"
2.
2
Xn are NID(&J" 0" )$ 0" known.
Find the C.,R lower bound for the variance of unbiased estimates of
jJ.)2
2: (\ _
L=
1.
(2 n 0")n/2
In L = K ""
2 )v
jJ.1)
- 78 -
.~
but
ax
*2
Example (Discrete):
x is a
hence
Binomial
x is B(n, p) -- find the C-R lower bound for the uJiased estimates at p.
L
n )
=( x
In L
pX (1 - p)
K + x 1n P + (n - x) In (1 '" p)
:or
din L. ~
op
.. ?;}
n-x
In L
p2
E.:!
-:-
I-p
x
=~
P
;:
['d J= ~
Hence
In L
} p2
i >- R\l-p)
~:
d~
but
(!)
YI
'\
\-f )
n-X
(1_p)2
.. E
Y- '..L~~)
- - 0()
n(l-~
(l-p) .
n .
p(l-p)
= n k2
= E(l-p)
Recall that the C-R bound is achieved if' and only if de!) is a linear function
"a'dlQ.f --
of
if
d~~:f
40:
Find the C-R lower bound tor the unbiased estimates of the Poisson
parametsr A. based on independent observations X:t" ~, ..., Xne
L = Pr(x)
In
L = In p
= pqx
+
x In (1 - p)
1
x
=--p
I-p
.. 79
E(x)
+
p(l-p)
= I-p
p
= n k
p (l-p)
thus
To find the M.L.E" of P8
1
x
=- - - o r
p
I-p
1\
p .., l+X
E [d(X~
or
= d(O)p
do +
+ d(l)pq + d(2)pq2 +
~q
+ d2q2 +
.00
c~o
+ d(n)pqn +
+ dnqn +
ooeo
fOO
5 1
=1
=0
~. == 0
p* = 1
if x.., 0
=0
i f x~l
ioe .., the decision as to the status of the whole lot is based on the first
observation"
This is unbiased, but o~o.o
(this example is used to show Why we don't always want an unbiased estimate)
i *' = pq~p2
p'
.' since this estimate is the only unbiased estimate by the uniqueness of power
series o
.J
co
de!) f(~" Q)
dx
==
or
for the discrete case.
These equations" however, are in general rather messy and difficult to handle"
Problem 41. f(x)
k xk- 1
1.
exte~sion
Multiparameter
We have
I:
fey 91 , 92"
'"
Assume the same regularity conditions as for Theorem 24 in all the Gts.
Denotes
>tl
""ooe
~k
..
..
A=
"
"
~l
. / \ is non-singular
..
"""
as in theorems 22, 24
Let d(!.) be an unbiased estimate of
BO
~(~)
that
"J.
E [d
81]
=1,'
==
O~'
= l\ ~
- 81 Theorem 2$,
where
1. = (1"
O~
000"
0)
k components
or
~2
~k
fle.
0
0
'1ca
ad:!.
"u 0..
~k
f)
(>
"
0
e..
Observe that
a i 8i
J2~ 0
k
ad:!.2
2 a1 "" ~ ~ a. a j A.
. 1 J=
. 1 J.
J.J
J.=
r(J<,!)
We wish to maximize<:V(a) with respect to a to get the best possible statement about
the bound of the var'iaii'ce..
k
(1) (!)
_ _ _ _T
~
.L--.J;l.:=~l
= 28- -
a J.. AiJ.. -
i"
a.a j X..
J.
J.J
'---'---=-------
'_
- 82 -
d~~CP
or
=2
2~>-.u
- 2 ~
.. 2
J=
a j "\A' j
~
j=l
or
A..
~J
=1
~ a.
j=l
Aij
=0
a'
j=l j
A:t. =1
J
~ a~ ~ .
j=l
where a
= 0
= 2, 3~ OQO~
i=J.
a?
J.
~ a~ A. J~
j=l
J.J
=~~
;max
Therefore
(~)
aO
0 A.
i~ j. i a j
= 2 a~ - a~ = a~
''ij
- 83 -
or~
as stated in the
O"~
v
I1
theorem~
>'22
C$$
c-
"
Pel
"
\:2
D1
o_1t
.1:.
An
0
III
E [SiSj
8.
1.
Examplei
00.
d 1n L
=-e"
A:Lk
~
Aij
--
7'1':k
"
<>
Remembers
~k
-ax b-l
ab
:rex) = - 8
x
reb)
x>O
=n
In a - n 1n
nb
III -
denote
82
= .~lb L. = n
Au=
A_
."2 2
r(b)
2
E [ 81
r (b) - ax
+ n(b...1) 1 n x
..,
-
In a .. n
fo [
1ll r (b )]
J.. F2 (b)
-1 = -
= E [822 J
:I
'd 2
J.n L
In L
nb
da"2-- "";~
2
d b 2-
= n F~ (b)
+ n 1n x
- 84 -
-1
F~{b)
a F~(b)
2
=--~;;...---
n [b F~ (b) - 1
- ~)
1
Note:
See the
e~'l"lJ.ple
J.
- 91
- ~ a.
i=l ).
_2
s. ]
).
~ o.
[d:J. - 91
~
1
a i 8i ]
function of
'i.
is essentially
is a linear
~ a. 8.
i=l ). J.
and, as before, the M.. L\OE. (corrected for bias :if necessary) attains the
lower bound (IF there is any estimate that attains that lowe!' bound).
Problem 42:
P:r
X:t"
QUI
[x = x ] =
X -
l' ..
1\
) pr qX
p +q
=1
1.
If
l'
is known
2.
If
l'
is unknown
85
Problem 42 (b):
Show that it I
with density
I A.
f1b)
~-------~--~------~~--~----~--------~-------
4.
Properties of m.l.e"
10
linear function of
2.
Jo
d~nJ=.
other
4& Among the class of unbiased estimates, then the m.l.e. are b.atln., otherwise
(ioe., if the m.l.e. are biased) we can not say that the m.l.eo are necessar
b.a.n.
ref: Le Cam; Univ. of Calif. Publications in Statistics;
Vol. I, No,) 11, 19.$3
-- The class of efficient (Cramer) estimates is contained in the class of m.vou.e. -which is the real reason we are interested in attaining the CaR lower bound.
-- 1, 2 are finite results, i"e. for any no
- 3, 4 are finite (asymptotic) results -- for large
n only.
.. 86 -
Eo
eno
1:11
if the distribution of
is (mathematically) independent of
that iS i f in
Q
.)
h(xlT)
~v
llt~!!:
..
1. n
(.[2n)
CJ
Xn are N(~, 1)
~::>
~x is N(x,9 1)
~Xi) ~
Tfii ~
"'------v---------""
heX IT)
X given
- 87 -
~Xi
~ Xi'
IJ!1ltinomia1 and T
=~ xi
n i
Ao
is sufficient for
NO(J> :3 -- Normal
IJ., ,,2
x;
are
~ (xi _ i)2
1
Problem 4:3~
Find sufficient statistics (if any) for the parameters of the following
distributions:
a--
a . e-ax xb-1
f(x) =
Gamma
x>o
((b)
b--
power
f(x)
=k
C"
Beta
f (x )
III
d..-
Cauchy
f(x)
x k- 1
pta,1
O~x~l
xa ...1
b)
=; (
O~~l
2)
1 +(x .. lJ.)
e-
nego binomial
p(x) ...
f--
normal
mean:
.,(, +
variance;
(1.. x)b-1
+ r 1'" 1) pr qX
r-
~ t
mown
,,2
--------------------------~----------------
~~:
[X - 0)2 ~
ETfEx(XLT ) _
~2
..
- 88 Proof'~
E [x21 t]
[E(X t)]2
~ F~[E(XI T)J2
E.r[EX (x2fT)]
we get
replacing X by (X - c)
tx -
cJ2 ~ ET[Ex(X
Proof:
coo
c f T~2
ET[Ex (X/T)
co
cJ2
is independent
2-
O'~
..
E[d(!) -
= g(~)
... E@(!)]
g(~IT2 ~
III
ET['t'(T) -
g(~)
.. c
g(~)J2
~ 0'2
/' 'r
" : .. Of
.. 89 Examples:
1....
Binomial
Xi is B(l, p)
1:1
1, 2,
.,
x ) pI(l _ p)n-X
n
OJ
where X
xi
1
(since f(x) is ordered, the coefficient
term is omitted)
JS.
'C1
xl
is an unbiased estimate of
pel ... p)
eo
in n trials we have
PI{Xr
Pr[Xl
Pr[Xl
1:1
success I X]
= olx
I:
llx]
X successes
II!
X/n
-1 .. -X
"" n
therefore
44:
...
Negative binomial:
show that
10
2B
find E[Xli
Example No, 3:
1 if SUCCEUSS on trial 1
0 if failure on trial 1
II
II
2Problem
p:>
xl"
-Xn
lmown
X is sufficient for p.
(1 = Xl) is an unbiased estimate of p,
nO"iiegXi == 1 if faUure on trial i
0 otherwise
..
(0, s)
Uniform distribution on
Z ... max(Xl ,
E[X i ] .. ~
'f (Z)
II
x2'
01
so that
E[2Xil ZJ
Xn)
2X
II
~Xi
is sufficient for
So
is an unbiased estimate of 9.
or>
proof~
90 -
uniform on (O~
Z then
J 2( ~ ) ... Zc
zJ
I:
...
is
Z),
then E [2x1IXi .c Z
Thus:
Xi
J 2Z ..
E[ 2xi Z
~l Z + ~ 2Z (1 - ~) Z + ~ Z
(1 + 1) z ... ( IL,+ 1 ) Z
n
n
T'" ~
Xi is sufficient for
~
we know that
A and
.J.
(~;"j
(see example No
~]
e- A
PrLx 0]
..2
so that
is an unbiased estimate of
.. 0
if X.J.' :> 0
then nO"
~Y.
1
1.
e-'A.
2 on po 86 )
E[
e-'A. ..
91III
III
.\j..J
III
1 -
n1 )T
r.
1 Ln(
Ii
III
T 0
1 )T]
Ii
is unbiased:
CD
E ['reT)]
1 ..
(1 - l)T
-n
( 1 _ ~ )T e -n>..
n
(n>..)T
T :
45:
III
e ->..
--~------------~---------------------------
Poisson
f(x) ... (
~xi I ~f
is sufficient; but
{2n
!n = X
is equally goodo
)n
""~
however
~Xi
92 -
Example
4: Same as No o 3, but
then
T...
~ (Xi
is known"
lJ.
"" lJ.)2
is sufficient for
cr2 ~
-~-------~~~-~~----~-~~~-~--~-------------
Remark:
since it is
independent of
, Theorem 27:
Proof:
-=
h(T)
O~
=9 denotes identically
equal in
dl (T)
E [d(T}]
Q
II
g(Q)
., 93 Eg[t\(T)]
g(Q)
J..
0 for all
Q"
or that
Further -- let d i Qf}
that E[diQi T)]..
holding only if
o~ ~ O~t
is unbiased and
is a funotion of To
dv
'f (T)
III
d(T)~
Example No"
1:
X is
Binomial
B(n, p)
x .. 0
heX) ( ~ ) px (l...p)n..,x
~a
p
======~~> heX) .. 0
n"
For this to be identically zero in p implies that all coefficients are zero
which means h(X)" 0 for every X" 0" l~ 2, ~ 0 ~~ n,
therefore
heX)
hence
Example No.
2~
pta
~ is m.vouoe e
Poisson
T
III
~Xi
is sufficient
~
1;1
=0
1.
implying h(T) .. 0
on the integers
- 94 CD
or
h(T) (nA)T
T 0
TJ
1\
3~
Example No o
Normal
Xl' X2 ",
<)
.,
T
n
[h(!}
N(~, 1)0
Xn are
N(lL'~)
e~ ~]e-nX1L eli
~s
CI
IL 0
=-
III
therefore
If
Xl' X2' ~
complete for
Remark~
(J.J
X is completeo
Xn are
0",
2
N(lL, a)
(..X, s)
2
are
a2 c
g( lLJI aC L
of
In particular, if'
2
2
g(lL, a ) ... lJ.2 + a 2 "" E[X ] ,
!n ~X2i
is an unbiased estimate of
~ ~X~
2
-2
.. (noel)s + nX
n
J.
= 0,
A,9
is m.v.uoe. of
is sufficient and is
-fin- .- '!f S
or h(T)
1J.2+ a 2
1 ~ 2
so ii A::d Xi is me V 0 u. e.
of
2
2
IJ. + a
- 95 !:oblem 46;
."
are
Y1 " Y2,
~p
In are
Xn are NIO(~I a )
N(~.t
a)
~(<; a2 )
,
Sufficient statistios ar-e
2
(X$ Y', sx'
s;>
~
(i~e","
1S.,
X2 'p
prob~e~
00
g(F)
IS
)X dF(x)
a:
E[X]
provided E(X]
<.
co
-co
g(F)
F(x)
g(F)
g(F)
1:1
F(b)
""
~
Pr[x ~a1
F(a)
(Xl" Xn) such that F(Xn ) - F(Xl ) ~ 1 (two-sided toleranoe limit problem)
co<.
., 96 Theorem 28g
If
d(Xl ; X2.l1
E&qp J =
Proofg
()
0,
g(F)
JS."
In) is symmetrio in
then de!)
Suffioient statistic is T
is mov.u.eo
= (~Xi" ~xi"
X2"
of
'I
0$
<l
<)
0$
Xn and
g(F)"
~X~)e
..
(I
nJ
solutions 0
Assume (as is true with. probability 1) that all the I's are distinot -'" :1:f.'
x_, X2S 0 0 "I XIt is a solution, so is ar.:y permutation of the XiS"
-j,
There are nJ
solutionsQ
Xl,
==)~>
g(T) .. 0 "
,
i
... 97
where:
p(tl' t 2J
.,
..~ Qjt
CD
Hence, if
g(!.) e
jUll
p(~)
-CD
by the same type of theory as :in the normal case (uniqueness of the bilateral
Laplace transform) this implies
or g(l)'" 0
Example 1:
EstiD:ate E(X).
Xi,
X2, 0 0 .~ In'
further, it is unbiased so that by theorem 28 it is mj)v.u.e.
e
Example 2:
is symetric function of
Estimate
F(x).
l..f------~-----~--:---===--:.:::::,,=-,..-.i;'-'~derlying
population dof.
sample d
-<.-:",~\~
01
.\1:::;> ..<'~"
. ?
j'
Fn(x)
III
where
~1-' (x)
III
if
Xi(X
]..
n
F (x)
n
= .n i=l
~
Wi (x)
Ii
It then remains to show that it is unbiased$ i.ee, for each x: E[Fn (x)
~.
J . . F(x)
For each fixed x the number of x 'Ex is a binomial random variable with
parameters [n, F(x) J-- and henc~, since the sample frequency is an unbiased
estimate of the binomial parameter, we have the required unbiasedness.
Remark~ Fn(x)
) F(x)
p
for all x
Ii
.
98
Problem
Pr[X '! a]
(a known)
with dof" F(x)o
~xi
liS
is sufficient for
(12,
(l
a)
b)
of the form as + b?
(12
~----~---~----~------~~_._-----------------
Fit x2.estimation
.- most. generally applied to multinomial
~- us~ associated with Karl Pearson~
ref ~
Multinomial
situations~
_.
58
Dec 0
Distribution~
with~
n experiments result
in:
v1 v 2
~Pi
I;>
vk
~v
... n
,
1.
Characteristic"function~
~
T
vl' v.2' c , v k
(t1 , t 2,
.,
t k ) .. ~
(I1.
v1
:l
n
v 2,
t
e
e 1 +
u.
0,I
0".
~~
(Ple)
Ge.
(p e k,
k
- 99 -
(t 11 0,
vl' v 2 OU$ v k
.00'
0) "" (Pie 1 P2 +
II
[l1.et l + (1 -
GO.
+ Pk)n
I1.~ n
Asymptotic
a)
b)
distributions~
""
vi-nPi
--~ nP (l-P )
is AN(O" 1) as n --.o.J>)' CD
k
k
2
~ 2
~ (vi'",np.)
~z. = A ~. has a
i=l].
i-l
nPi
.. ')
x""-distributionwith k-l d.f o as
.
v l' v "
2
~ ~
parameters
a,
n-7(Q,
k dependent to
h:t.t
~~
I)
------------------------------------------Example:
(Homogeneity)
We have 3 sets of multinomial trials with possible outcomes Ep E2" E )
3
with probabilities 91' 9 2,91 - 91 - Q2 for each set of trials.
v l .3
n..t
22
v 2.3
v.3l
v
32
v3 .3
vol
v. 2
v.3
12
21
In L = In[n.. 1
V ll p v12
-~
12
n.. V31.3 p v 2l p v 22
.~
21
22
P v 23
23
equations~
91 + vol Q2
(vol + v c )
v l2 Q1 + ( v $2
v 03)
-I-
Q2
vol
v 02
N Q1 + N Q2
v"l
+ v 82 = N -
v 0.3;.;
v Q2
Ql
V 02 Q
v 2( v 01+ va 2)
N
p v31
31
p v 32
32
P v337 + 1n K
33
- 101 -
-----~----------------------------------~-
!!oblem 49:
( Independence)
where
i=l
Given:--s
p.
1 ;
IlS
j=l
J.
1:3
1 "
general case
serie~
p ..
't"j
Pi" 'l;'j
of n
_The probability of Ej
k
j=l
events
x2.estimation;
rc
=1
i th series is p..
J.J J
J.J
-the PiJ'
G,
v ik'
are functions of
, Jl:ll
vij
= ni
I'jlf
ij ]
= ni Pij
-. the use of
s trials as one
.. 102 -
maximum likelihood
2
m:iriimum X
20
transformed minimum X2
4$
1(1
Maximize
k
~
Q;
p~
:'.J -
'9
" -
(vi - niPi)
,J
ni Pij
or to solye
Hence, equation
r2] is the
eql~ls
>0
as
0
N ~oo
same as equation
[1] except
that under suitatle regularit,y condiGions the solutions of equation [2] tend in
Pij 1
for
i 1, 2, , s.
X ... Xm
III
...
R.
are
2
) 0 faster than does X --....aS~l!l.pi.;otioal1y equivalent.
p
therefore methods
(2)
and
(3)
~~
40
Transformed
ni
(V";j.:'ni~ij)
'''.,....,'1
1I'~i.'n''!!LL
Recall that 1
In(Xn..~)
-b;-
g(X )
n
is
cf[
S(X )-g(~)
Cng'(~)
g t (IJ.>] 2
is ll.N.
vii
qij .~
x2
X; ~ ~
i
32..
~~
i
D1
[g(q~)
g(P~W
Qij Lg.(qij>]
is simple.
Summary:
.
ni [g(Qij) ..
qij [gl(qij)]
Method (2)
g~;ij) ]
Pij
products, i.e., Pij Pi'Cj
2
[minimum x ] -- most useful it Pij
can be expressed as
as in problem
49~
can be expressed as a
J-
2
M9thod (3) [modified X
Method (4)
same as (2).
\) t ,1----1
t V'\\.
01-) / V\,
\) '2'
- 10.5 -
prooedures:
1. Linear Trend:
-vobs.
p.)
ni 2
total
prob~
Pup-A
12
P12- l - p + A
v2l
P2l P
22
P22 1 - P
Il
31
31
., P + A
v\,
v
32
P32
III
1 .. P - A
-t ~~2.
~ [vll"~(P-A)J +
where
a 2 [v 2l-n2P] + a3
rV31.. n3(P+A) ]
a i ni'...L +..L)
Vil
i2
These tw equations yield the following two equations which can be solved simultaneously for p and A ~
Note:
remember that this procedure is asymptotically equivalent to the m.l.e o - therefore the asymptotic variance-covariance matrix can be found by evaluatin~
!.2 ).2y 2:
~#
! ~2y2
2U'
l_d_
- 106 -1
a..
. J. = n1 ( v
l)
+
v12
ll
Recall
E[vlIJ
= ~ (p-A)
1
1
I
--p--). ( p-A + l ..p+l ).. {p..A) (i.. p+A)'
I
p 7 p(l..p)
Similarly~
...
-----------------------------------~-----~-
2e
= general
xl'
~$
ni
=2
xs
dosages
0,
ns
receive dosages
09
va
o~
nI" n 2"
vI,\l v 2'
:>
e-(-<+~xi)
= l+e.(-<+~xi)
1 .. Pi
Vi
q.J. = -n
107
2
Applying the transformed- X method to estimate
g, g',
.,/."
we have
,,2 ~ n.
i=l
---------~-----------~----------~---------
Problem
50~
Consider rc sequenoes of
where in trial (ij )
i
j
"/"i"
10
maximum likelihood..
20
minimum modified X2
~j
by
1,,2, 00 ."r
1,2,.06,0
- 108 Problem as stated produce.s r+c equations in r+c unknowns" but their matrix
is singular -- i.e., the equations are not independent by virtue of the fact that
the swn of the I' equations for the .,(i equals the swn of the c equations for
the ~jfl
Therefore, reduce the number of sequences to
and add the restriction that
P22
CI
4i
ioe o " reparameterize and find the equations necessary to solve for the
new parameters<ll
Given
51~
Problem
s
sequences of n i
maximum likelihood"
20
2
minimum lOOdified X "
A by
1.
-Minimax estimation:
Recall that
Q]
R[d, Q] E[d(!> -
58
g<.~) if
is minimized by dqf)
g(~r
[see def e 24 on p.
54]
Cramer-Rao
cl;:"
[l+b'(Q
d ;;;..-
nE[d ln
E[d{~)
g(~)J2
x ~1
J:.:l+ b'~lL
nE
~(!) -
III
<{ +
Rld,9] >-
,/
f1+b (9)1
nEI~ In
-an92f(X~J
[. d
E[d{!)] + E[de!)] -
Lb(9)}2
+ b 2 (Q)
k(Q)
f!&12
J
dQ
If we have an est:iJnate
d(X) which actually has the risk
k(9)
estimator with risk less than k(Q) leads to a contradiction.
Example:
~"X21 ~
0,
J2
g{~)
Xn are N(~,
J)
2
and we know R(x). ....2n
~0
This method based on the following theoJ:em which will be stated without proof G
~eorem 29:
!.f ~ Bayes estimator [deft 23] has constant risk [defo 25] then
1.t 1.S minimax/}
d(~)
de!)
ref:
Lehmann -- section
Example:
4,
is constant.
[d(!.) ..
gee) ]
f(x, Q) dG(9)
Q.
p. 19-21
Binomial -- X is binomial
B(n p p)e
d(X)
rn
!. +
1+.Jn n
1
2(1+Jl}
54].
If'
.f'n"
f(p)
then d(X)
I:
K P2"
12. 1
(1 _ p) 2
is Bayes;
rn_!+
l+Jii n
1
2(1+$)
is
1
rn-- - -- - - - -.---.-,.
.It.::;
o ;!.
-+-
Problem 52:
Compare
( d
Ho
Efd (X)_p,2
l - 'j
Rd' (p)
and Rd (p)
m
for
I:
:e(l-p)
n
n =25..
X
- ; d = minimax estimate)
n
m
.~
Q).
e.go:
(F, G) =
sup
1F(x)
ro (F,
- G(X) \
.co<x<oo
Ii
(F. G)
(fr(X) -
G(x>]2
<lex) ; G(x)
(see p. 97).,
G) e
.. III -
Q'
is a minimum distance
minimized by choosing 9
Note:
es~mate of
= 90
if
f. fF(x,
t
Q), F (x)] is
n
We want the whole sample dof. to agree 'With the whole theoretical
distribution ... not just the means or the variances agreeingo
~
which yields
min
Remar!.:
Proof:
"..-
n) 9
differs from 9 by
15
jJ
sup
is that estimate of
eo
is a consistent estimate of
I F(x, ~)
.. F(x" 90 >J
> 15
"'CD<X <00
Let x'
F (x) - F(x,
n
I"
therefore F(x',
thus
< :315
sup F (x I, Q) - Fn (x W>J
1
15
>j
)
is consistent.
",
15
sup
minimizes
Fn(x) .. F(x"
o)
Q)}
N(IJ., 1.
~ = 1
p
Fn(X)
_::::;:::....,,==....--
2/3
1/3
.\
o
Note:
(x)~
n.
To find the minimum distance estimate of IJ. an itterative procedure was used$! whicJh
started by guessing at IJ. and then finding F(x, IJ.} at each X.o
J.
204 zi=xi-IJ.
P(zi)
767
lilt
-1.4
"",0.4
1,,6
..
,,081
0345
0945
zi
...1.,3
-0 0 3
10 7
f(zi)
. . 097
0382
~955
~288
2 e 33
P(zi)
..371
&952
299
20 29
f(zi) ;:
0386
.956
284
2,,3
Other values of
therefore
~~em
e.
(\33
X
3
1 00
Fn(x)
X2
Xl
IJ. were
IV
tried~
(x
..
(.(,1- 3~ b)
284,
=2,,33)
IJ.p
113 Q.hapter
I~
V~
Basic concepts..
Xl' X '
2
Der Q
~, X
The hypothesis
Alternative
specifies that Fe
33:
A test is a function
iQe.~
if f(!) ... 1
7- f{
H with probability
... k
II
tI
II
II
it
Illl
34:
.,( if E fI(!) ]
A test is of size
fr'(!lJ
rf(~) dF(~)
for Fe
? ~ .,(
"'00
~~:
~...12~
Power of a test is
r~1
E[1'(!)]
Fe
1",. ~
and is denoted
~96 (F)
Def e ~:
for
H when it is true
as
Fe
~f (F)
7- (
n ~ (I)
Cfn~
1: co<!)~)
~ co<)
.. 114 ..
l-~
+-----/
I--
--Io-;-
7~
-- 1: 10)
Def. 38;
Let
p(
Let
fn and 9ii
10
/\A.-'
lim
N(
--
~ to
~ is defined to be
~'1," 7~ ""'3 ~)
---~-----
p-';'O N(
(0' mown)
alternative~
IJ.
H~
10
t he mean te st ~
2"
reject
H if
>0
... ~-z
X)'_..-
(J
Vn
[use the large sample distribution of the median, i"e-the sample median is asymptotically normal with mean
... the population median and variance = Ra2/ 2n
if test:
a)
rejeot
'7 "1--<~
~!
- 11S b)
find the
A~RoEo"
ieee
lim N~f',
IJ.' -7 0 N(
II g
r;
,*U 2
.,(,9
IJ. r;
co<."
@)
(3)
Ao
I:
Test:
Ho : \
I:
the number of
H~
<
-
A0
(q ~p) at level
co<.
JC]." x 2"
I:
pJ
~: Aq
= "'0
/)
"
oa X
X is B(n" p)
~::!!lple:
"'0
.. infx(l(x)
A.
under
Problem
-.
..
H if.
tJ np(l
- p)
;;
zl..<./
.., 2
SS: Compute the power of the test (using the normal approximation) as a
function of q for n"'" 100 and p == 005
for paired comparisons" to test if the median is equal to zero set up the
series of observations dl , d 2"
0"
dn,
di
= Xi-Y i
- 116 This test that the median is zero is often referred to as the sign test, sine(
in this case is merely the number of d with a negative signa
i
pete 40:
>-p"
=P
Theorem 30:
\>
where
Proof:
ref:
Pr
l.L -= [np]
. np
Cramer po ,368
[Xp ~x J
III
( j ) [F(X)] j
[1 - F(X)]n- ..
j-l.L+l
li'(~)
To get the density hex) [assuming that F(x) has a density f{X)]
we differentiate the summation, getting the following two summations
[trom each part of the produot thatoocurs in each term summed]
n
hex)
-=. ~ ( j ) j ~(x)]j-l
J=l.L+l
[1 . . F(X>]n- j
... ~ ( j )
(n - j) [F(x)]j
j-",,+l
f(x)
[1 - F(x8n-j-lf(x)
The corresponding terms [in ,a(l .. F)bJ cancel each other except for
the first (j'" l.L + 1) term in the first summation which has no
corresponding term in the second sum.,
/\
\>
.
and variance . 2 1
f{\)
p(l..,p)
n
is asymptotically
- 117 ~ob1em
56: Let
J$;
ha11e density
co
) rex)
where
dx .. 1
""co
and
rex)
....1
c
f(
X"/J: )
C1
co
x rex)
dx .. 0
x 2f(x) dx -= 1
"00
is symmetric about Oc
Find the condition on rex) such that the AcR.E. of the median to the mean
is greater than 1" Use large sample normal approximations tor both tests"
Specialize this r.esult to
I x..,u I
o;-~ ..
f(x)" ~ e-
CJ.)
<x <0:>
--~-------------------~-------------~-----
III.
One Sample
H:
o
Tes~~
F .. F0
completely specified
~ \)
-J..
:Ell:> F == F.,,..F
.I.
0
problems:
1)
2)
ilslippage test!i.
is usually not
4)
re~
Randomness in Time:
Assume:
to
t)] e
-At J.~t)n
n'
At
"-.
GOO
~ t.
j=l
Pr[ti>t
tet)
III
= density
f(t 1s t ,
2
I)
T1
== t
At
~ e "A~ti
0 <u;
Tn
t ) ..
e"'
s:L"lce the
IS
are independent
is obtained by transformation
~ ~, Tn ; n)
= density
ot Tl , T2,
!>
.,
Tn
0'
..
,\n
-AT
1\ e
ten)
0,
I , I 2"
l
.,
If the XQs
between the
Xrs
Q)
00
nl
f'(Yn ) dyn
Yn""l
Ii
f(Yi+l) dYi+l
Yi
Y3
f(Yi...l) dYi_l
/)
"'00
f(y.)
J.
Y2
f(Y2) dY2)
f(Yl) dYl
-00-00
note: ...- the observations above Yi are constrained by the next lower
observation$ since this is an ordered sequence
-- the observations below Yi are constrained from above
integrating this we get
gi(y)
20
lllI
By differentiation:
Pr\!i.( Y] =
Prei
1:1
J=i
( j ) [F(y)]j
[1 .. F(y)l n- j
:.J
... 120
3(
tn.i)~ti.ln
[F(y)]i-l
.., F(y)]n...i
fey)
_ _ _ _-"'~1!i.._;
1i
Y
(1-1) observations
in the interval
(-eo, y)
(n...i) observations
in the interval
(y" + 00)
Example:
F(y)
=Y
n).
J.-,l
gi ( y ) = "{ii..n trI:IYl y
(1
.. y
)n..."i
81 " Yi-Y i - 1
i 1 , 2,
QI
n+l
n+l
i-1
8 1 -1
Y .. 0
Problem
b)
find
1)
{~5~]
n+l
11) E [
11
1~ r51 - n~~
----~---------------~----~-~--------------
-121Lemma~
Probability
Transformation~
PrI!(X)~U]
Pr[X~F-l(u)
F'"'l (u)
...r.-
_ _ _ _ _ _ _.......I - -
in! [F{X)
F~"'l(u)J
au
One Sample Problem can be put in the following form by use of the probability
t ransl'ormation:
Given Xl' X21
')
0);
In; let
Ui
III
Xll)'~' I(2)~
H:
o
i = 1,9 2,
Fo(I(i
U1 , U2s ..
eo...
!9
Un on the
and under HI ~
has distribution F
1
two-sided problem
[F~l
a Go
one-sided problem
G{U)~ u -- with the
n which is
AoN~(~,
rk )
ii is "too small"
.. 12.2 -
n
20
-2
i=l
In U
i
e~
....
-""
consistent
for the one-sided problem
used in combination problems
reject H if' X2 is Of too largeU
30
<.02
i-l
=.
4G Distanoe
NOa
Type Problem
n-...
sup [ u - F (u) ]
sup
III
o~uu
, Fn(u) - u I
o~u~l
.- is consistent
5a
sup
a~u!l
III
sup
a~u~l
-----
Fn(u)...u
u
one""sided problem
tFn(u)-u)
u
two-sided problem
6.,
12.3 -
eu~
)[Fn(U)
....
__
---
ul du
n+1
7~
wn = ~ i=l
~'8i ~ I
co
n+l
8f)
~ = ~ S~
i-1
J.
a_
90 Ul
., 1
is AoNo(O$ 1)
-- one-sided problem
generally not consistent
Problem
58~
Ul
~co
G(u)
:I
G(u)")u
uk
O<k<l
10. ,,2
11$ NeymanYs smooth tests
-- discovered by Neyman about 1937, but never generally used
-- re.t'g Neyman; Skandinavisk Aktuarietidskrift; 1937
Pearson ... Biometrika; 1938
David"" Biometrika; 1938.
- 124 -.
Problem ,9~
Xl" X2,
let
o.
= max
Xn
0,
Xi -
min Xi
l~i'!tn
l~i~n
a)
b)
Suppose F(x)
R...
E(R 1- k(f"
show that
(J
x 2(>:) dx
)f(X) dx 1
where
1: f( x""tL
a
~1
n)O'o
-------~----------------------------------
Theorem 31:
F
If
!!22!.~
is continuous,
Fn
- Pr
[I Fn (x)
.. F(x)
I <' s
p>F for
8, (3
for all
au
Xv
we oan find
x]., 1 -
n '/ N
(3
F[x(i)
= X(o)
no
Let
has a
,,2
i=o
We oan find an
1>1
such that
n.
1.
=;/-{x.
n J
distribution with
k+l dof"
i~ interval
- 12,5 -
or
Pr
[X~l
MJ /' 1.. S
<.
Since the above sum is less than Mj each term and aJ. so its square root are certainly
less than MJ therefore~
Ini..nPii
PI'
[ -JnPi ... Mfor each i 0, 1, , k + >1. 6
<
Recall that
~ n.
j=o J
n
Now
J Fn(X(i)J
j=o
(k+2)MJP!
Choose n
rn
so large that
< 2'
J=O
JUO
for all i
11
<.
(i+l)M.JP"i
vn
<:.
<.
*
F[X(i)] - F[x(i)]
+ F[X(i_l)l Fn[x(i)l
F(x) Fn(x)
~ F[X(i)J
<C 8
8
'2'+~&'lS
with probability 1 - S
x'elalllonsblps:
.. 126 ..
..
~ <: F[X(i)l
~ < F(x(i_l)l
.., Fn[X(i)]
..
<~
F(x(i~ <0
Example:
Referring to test No o
4 given
on page 1220
If
If
G is true,
IFn(U). G(u)
1<
eo
sufficiently large
and thus
sup
IF
I Fn(u)
(u) - u
DO
I Fn(u) -
u, ..(&0
uJ '>s.
will reject
with
1) Xl' X2,
()
Xl' X2,
2)
note:
Xn uniform on (0, 1)
~1 Xn with exponential density rex)
0$
=a
e- ax
---~-~----------------~------------------~
Remark:
The Kolmogorov statistics, Dn, D+, and D- are in fact invariant under
the probability integral transform"
Proof g
sup
f F (x) "" F(x)
-a><x..(oo n
III
sup
FJF(X)] ...
o !F(x)~
S!
Fex)
sup
IF
0 ~u ~l
(u) - u ,
- 127 -
Fn (x)
JS.,
.i
n
Therefore:
III
o ~u-{l n
sup
, F (x)
..c:o~x <con
F(x)
D~
Distributions of
a) lim . Pr
n~CD
.,
where
, F (u) ,. u I
sup
rJ
(l
III
[rn D: <z J ..
_,
QI
e_2z 2
(due to Smirnov)
0 {z <co
III
lim PrrJii D+
n-7 co l~
n
.c. z
J.
22
CD
Ma.th~
Stat"" 1948
The simplest proof of both results is due to Doob in the Al".nals of Mith"
Stat~"
1949
Some tables on the finite distribution of D are given by Massey in JABA" 1949"
ppe 68-77Q
n
Consider now that
n- test is to reject
H if
n'"n
> en
where
lIn. .,( I
2n
~
- 128 Suppose
is ture so that
maximum / u - G(u)
Then:
Pr[D::> fin
A at u
1~
~
~
Fn
III
l)
from
G( u) with
Uo G
Pr[uo - Fn(uo )
> en ]
Pr[Fn(Uo ) - uo<-e n ]
Pr[nFn(uo ) t::n(uo - en>]
Prob1em.Q.:
Using the normal approximation given an explicit f.orm for this bound
in terms of
x
2
,J, (x) = .1n, -<, 'I
e _t / 2 dt
J2n
Test No 0
IlC
5; Renyi Statistic
+
R
R
=0
Fn(U)-U
sup
a~u!l
= sup
u
Fn(U)-U.l
a<u~l
.....
Limiting distribution:
lim Pr [
n"-1 oo
rn R+<z]
CJ~
I
1
1
sa
[Fn(U)
lF~(~)
UY
co
du
-2uFn (u) + u
j
2
du
- 129 ttl
i
n
fo
recall:
un+1" 1
n+1
.. i=l
~
i ( u2+'"" u 2) + !3
-n
i
i
2 "'"
( -i )2 (u 1 .. u. ) ...,:r...c::J
1
n
+
1
~ 1-1
n
.. 1 +
2
u du
~ ~ (1
n~ 1-1
.., 2 )u .. ~ (~
i i
n 1=1
u~)
.., 1 +
i
J
.:F.,2
'n
.""4J
III
1 + 1 ~ [u 12n2 n 1-1
1
z]
~
2n
n~co
StatQ~
19$2, P3 20611
...R-.
0026
.11$
,,27
036
&7$
-.1F
02
03
04
S
J....
,,76
081
,,89
,,92
.98
-06F
07
08
09
1 0
0
.. 130 ..
Under the basic J::wpothesis the
Alternative tests:
acoept H
P OeS7
21.'
D:. F
sup D+
OQ08S
10
34> D. eS88
"
0088
J~r ~
0e964
067
BtU] under
H if'
Ho~
(,,75 - 040)
[rn z ]
Pr
[D)~ ]
no
(II
at
L(z)
mal
L(035)
Example~
-
In the following-
table~
{2n
)
..co
dt = PI'
f J
Z
<Xi
deviat~so
z <zco(
- 131 -
042
-0038
-0002
-0 0 02
.49
088
,,08
.53
040
009
.54
031
064
'?40
066
e08
042
.66
1012
088
011
-Oe38
1;p12
081
031
1016
&96
X(1)
1.16
009
Under H:
Ui
Xi
.35
8i
~
-;"09
.14
009
&04
009
~01
~09
.10
t>09
.02
009
,,00
1109
.05
:>09
c16
~09
009
t}09
004
009
n--!
Is i . n+l
.1.../
Ill;
Oe38
Ignoring the slight negative oorrelation between the S., one could use a
normal approximation with:
1.
E[W ]
n
l:
! .
e
0037
v(
W)
n
2e-?2 (0.07)2
lDe
~amp1e:
transformation:
H:t ~
u..
2
the X' s are X(,5) then you should use the probability
57~i?5)
It ~
X
t 2 i-l
r e- / dt
t'
- 132 ...
Combination
-
Probabilit1es~
- --
of Test
...
Ho:
U is unito rm
H:J.:
-2
ln Ui
Fearson~
-2
In (1 .. U )
2<:
tests~
i
..- found unsatisfaotory for moet applicationso
tests~
D,
~,.9
...
F - F0 completely specified
David~
Biometrika" 19389
x,
'0 versus Hl
oR, W~ , U)
F .. '1<' Fo.
timinimax ll in the sense that it has minimum power to piok up easyto-detect alternatives" maximum power to pick up hard-to-detect alternatives.
- 133 -
1:1
\!l
.,
II
9'
or
F>G
Hl :. F "G
In the parametric case one could use the normal approximation en d the two-sample
t-test on the means, but these hypotheses are somewhat wider~
All we need for these tests is to be able to order the observations; magnitudes
are not important; eeg. lv y. y \/~ \.'
\
~3~~W~
Test 1:
For the median test set up a 2x2 table classifying the observations
as above or below the median o! the combined sample. For example:
Below
Above
Total
X:
Y:
m+n
m+n
m +n
Total:
2
Test Statistics is the usual X for 2x2 contingency table s with one d.f,
Test
2:
If' H is true, then the Xis and yts are intermingled and the value of
o
r will be "large"; 1 Hl is true, then r will be "smalllt
- .134 Test j :
= 0
otherwise
m n
~ ~
0:
~j
i-1 j-l
Rf
where ~ = the sum of the ranks of ~ in the combined sample ordering from
the smallest.
Test
4:
2.,.
Prove: U = mn + m(m
1) ...
~ ~
n
0:
3.=1
~ R~ _ n(n + 1)
. 1 J
JD
= sup
\ Fm(x) - Gn (x)
-oo<x<co
Test
2:
.l:fin
_t2~
dt
-eo
a..)
let Q.
X=
:t~1 r(
RX
m + ~ ...
1)
I!l
8'
~) 0
~oo,
Ret lI' :
r
==
Mood Chap 16
fl\
Again~ there are two possibilities, i~ell starting with a run of l'S or of
yts, so starting with either, we must multiply the end result by 2 since
the two starts are symmetric.
Starting with the X's, they are divided into k groups, all non-zero. To
find the number of ways of doing this, consider the ccefficient of t m in the
expansion of (t + t 2 + t3 + ., )k.
(t + t
23
k
t
+ t .;. He) = (l-t)
==
=tk
k
(1 .. t)-
t k ( 1 + kt + (-k) (-k-l) t 2 +
=m -
k" and
_ fk + m - k - 1) 1 _ (m .. 1)
-m k)t (k .. I)l - k .. 1
i) (~ : i)
XiS
-136Therefore:
OOd Case:
E (r)
=m
~
+n
V( r )
= em
t'
2mn (2mn - m - n)
+ n) em + n _ i):;:::::;
N =m+n
Where:
+ 1 ;::;:;; 2 N..(
= No<.
2 2
N,.( ~
= N!3
0(
+ ~
=1
lim
N~
Test:
00
r
2..<13
H is rejected i f r
o
- 2N4.
is N(O,l)
rN
<
U=
~ ~
i
where
u ' ... 1 if X.
~
iJ .
=0
If H is true"
~(U) ...
..2
i
~ E(uij ) = ~
j
< Y.J
otherwise
-137. Pr
[y
> X~
= P
Pr
[Y > x
I X = x] dF(x)
-co
CD
= ) [1 -
G(x)] dF(x)
-00
E(tf)
=m n
E(U)
[{I -
G(x)}
I F(x)]
2 )
~4 E(u..
~J
::1
"
+-Z] ~
B(u.. , u. b)
E(uJ.' " u )
aJ
J
(3 )
E(u .. , u b)
~J
a
(4)
J.J
ijfb
(1)
+~
2i
.z
isla
+~
ifa
jfb
J.
2
E(U ) = p
ij
(4)
(2)
mn such terms
= E(Uij )
E(uab ) = p2
Xi' Yb
JS.:
> x,
>x
Pr [Y j
Yb
I~
x] dF(x)
-co
00
= )
[1 - G(xll dF(x)
-00
= B[ {I -
G(x)} 2 tFJ
(3 )
by similar argument
2
E(uij " u ) = ) F (y) dG(y)
aj
-00
=E
C; (y) IGJ
-1,38Thus~
, Exercise:
= E(l
_ G)2
==
IT (m + n + 1)
/~
(min+l)
formula to get all the moments of the distribution of U" then observing
that the limits of the moments" as m -+ (x)" n -7 (X) were the moments of
the normal distribution.
U -- may
=G
H F
Problem 65:
let a 1o' a 2 ,
let Xl'
~"
Define:
zi
(a)
IOU'
~-l
be fixed points o
= Fm(ai )
u.,
- Gn(ai )
F(a.)
J.
=f i
==
G.
-139(b)
=u
F(u)
=0
G(u)
"'_ 1 + u - 1 ... b
1 - 215
=1
In this problem let a
(c)
S. u
l-b
.s:. l-b
<.. u < 1
- - - - - - - - - - - - - -- - - - -- -
--
...
_-------
(b)
(a)
Y (5,5)
Fm(x)
-roj-1,-,_
..........
L!.
'I
l~:
--.J'
I
I
~>
Gn(y)
i
~
I
.r-
if
I
I
(0,0) ,.'"'-----~~-----
i.e., the sample could be plotted as
a two-dimensi ona1 random walk reaching
the point (5,5)--reject if the walk
strays beyond a line parallel to the
0
45 line"
Asymptotic distributions of D , if H is true:
ron
lim
m,n --7
1(z) has been tabled in the Annals of Math. Stat., 1948.
+
+
Also" D- have the same limiting distribution as D- (one-sample
mn
n
statistic) with the normalization
(see p. 127 )
1m:
-J.Lo- Test:
.
5):
where
u
(~) =...L. (
.;rn)
e-t /2 dt
-00
X is AN(O, : 1 Q)
Example:
where
2 i
r
(N+1)
i=l
Q =j
=m + n
XXXXyyYXyy
ItJ.
RX
i
11
11
11
11
11
111+1
(111+1)
= z .09
z .18
-.60,
.!. L '
when H is true p
= 0"
e,
i.e., P1 P2 " Pn
--+
O.
-141-
Test:
T -- reject H if Tn
> t n.,(
T -- reject H if Tl!-
>
Assumptions:
>0
(1)
-dd Ep (T )
.p
n
(2)
d Ep (T ) p=O
lim ~
n.
pm"
d
(3)
lim
n",,"*oo
ered
-dp
=c
J.'f
CJ
o
Ep (Tn)
Ep (Tn)
I p = pn
Ip = 0
l{
P =n
rn
=1
plJ"l (T)
n
CJ
(4)
lim
CJ;rr;J- = 1
n-1'OO
Theorem:
.
=..!. as
n{r1
The A.R.E. of
= (c*)2 =
T:~:'
"l!
to
lim
n-7oo
~~
is
00
is 1 -
~ (zO'( -
kc).
-142-
Proof:
Pr
>
Tn - EO (Tn)
t n:< EO
0"0
~J.
= ..<
0"0
lim
= z.,(O"O
EO(Tn)
n~co
= lim
n-1oo
0"
= Z.,(
(...9...)
0"
lim
n..-?oo
Pn
from assumption 4, as n
0"0
00,
( _..) ~
0"
kc
Pn
o < pIn
=Therefore:
n -1
kc
prlReject Hr
lim
< p.n
P~
=1
(z.. -
00
x-'"
kc)
d EP(Toll-)
n
where
c~l-
-d
= lim
n-j
. . , '"'OS
00
P= 0
to be the same p
n
= p*
or ---
y'ii1
k*
=---
0i
= k*c*.
1 -
(Zl_..< - kc) =-
AeR.E.
=( -r-C'
= lim
c ,)
Example:
n ....,
Coa2).(~d Ep
CD
(Tn)
dp Ff' (Tjt)
Ip = 0 )
Ip
Obtaining the A.R.E. for the Wilcoxen test versus the normal
mean test, i.e e
U versus Z
I-X
8'
co
p Pr
[Y
> X]
[1 - G(x) ] <IF(x)
-h>
- co
f (x - IJ,)
(x) dx
-00
-144~ E(u)
E(U) = mnp
= mn
(x) dx
IJ.=O
_ mn(m + n + 1)
Var (U) 12
E(Z)
x dF(x -
Xx dF(x)
IJ,J -
crV/ .!n + m
!
dE(Z)
dIJ.
~/1+1
vI
iii
ii
Var (Z)
mn
~
(J
in
==
) f2 (x)
-00
I-mn<m ; n + 1) ).
n~
oo\!
I ron
(] y' iii-+n
12
= 12
A.R.E. (U to Z)
;. [
=-L
Il1tcr
e-x2/2cr2 then
f2 (x) dx
2no-
1
0
.1
...........
V2n .j2
l /2 dy
e-
-00
and
12
J [ )
2cr/fl
00
(x) dx
12i
-00
12i
4ncf"
=1= 955
n
=~
All of which says that the A.R.E. of the Wilcoxen test to the Normal
(two-sample) test" if the underlying populations are normal, and we are
testing IIs1ippage of the mean" is 31n.
Problem 66:
(a)
=1
F(x)
=x
F (x - j.l.)
a<x
=x
<1
j.l.
a <x <CD
(b)
Test
1. Median
2/n
2. Runs
3.
3/n
4.
K-S
???
5.
~ 1" i.e,
= e-x
(2) f(x)
Remark;
a~ x
Consistency
yes, if the median
median of F
l
consistent for all
yes" if the median
median of F
l
consistent for all
of Fa
Fa = Fl
of F ,.
Fa = Fl
X
1
yes, if the median of Fa f
median of F
1
Robustnes s of a test (as propounded by Box) refers to the behavior of a
test when the various assumpti ons made .for the validity of the test are
not fulfilled.
Type 1 error - P.r [reject H when true under the assumptions]
Power
..4
r
- 146 -
1.
2.
3.
4.
Median Tests
X 2_Test with arbitrary groupings
Kruskal-Wallis Rank Order Test
Kolmogorov-Smirnov T,ype Tests
samples:
1.
..
2.
..
k.
1.
~l
\2
k
N= }"n.
J~ J
~~
...
.....
./
Above
mi
N/2
Below
n - m.J.
i
N/2
n
l
...
n.J.
nk
= n./2
J.
when N
a j +l ) define nij
- 147 -
=:
12
N(I~H)
where Ri = sum of the ranks of sample i taken wi thin the combined sample
ii
=:
Ri/n i
!to
Would involve drawing a step-function for each sample on the same graph.
Unfortunately nothing is now known about the distributions.
ref:
Consistenc"I:
The Median and H are consistent against all alternatives if at least one of
the sub-group medians differs from the others.
A.R.E. :
A.R.E. for slippage alternatives:
median test against the standard ANOVA test
2/rr
3/rr
H test
against ANOVA
1/3
1
.. 148 CRAnER VI
TESTING OF HYPGrHESES -- PARAMETRIC THEORY -
Cramer" ch. 35
Kendall" vol 0 2" chs. 26-27
Lehmann" IlTheory of Testing Bypotheses,lI, U. of Cale Bookstore
(notes-by C. Blyth)
Fraser" IINonparametric Methods in Statisticsll"ch. 5
Refsg
1.
't.-
POWER
Generalities
-- usually the XIS are independent with density f(x, 9), the density having a
specified parametric form with one or more unknOwnparameters.
For the parameter space.fl HO'
~. ~ e '--\
e CJo
0<
such that
=k
==0
where E [()
Power function'
Def'll ~:
~ (Q) ~ E
e Wo ]
~-<
[ I ~J
* is a unif'ormly most powerful. (uom.. p.) test of size ""-, if being an.
. other size cI.. test
'\}
2.
Ref 8
--
Probability Ratio
T~
Neyman-Pearson Theorem,
X is a continuous random variable with density f(x, 9)0
HOi
9 == 9
(simple
Hlll
0
~pothesis)
Assume!
f(x, 9 0 )
Assumes
f(x" 91 )
f{x, QoJ
>
OJ
f(x" Ql)
>0
9 == 9
1
(simple alternative)
for the same set So
- 149 ..
Theorem 33 (Neyman-Pearson):
The most powerful test of HO against ~ is given by *(x) defined as follows~
*(x)
1
elsewhere
- 0
against Hl :
&~ 0
Remark: This is what has been called the probability ratio test, since Neyman.-,
Pearson originally expressed the theorem that
I----+--~--
__~--~----
Proof:ltl
..<{k)
-o) .,
since
r:
Pr
[1f'O{x)
(x)
:.>
X has density f
lJ
oI
'-'(00) == 0
~(k)$
which is the
d.~
of
==-<.
- 1,50 2.
be
j/(x) t(x,
Consider!
!ll) dx
[(x) - lI(x)]
when f
- k f
follows since
>0" then'!t
- k f ~ Os then
1.9 and
rf! ... ~O
=.0" and
ff .. ~O
=:
)1 t
l dx
fllt
l!f
but
j*
Therefore;
and
l dx
f O dx
dx
::&.
- k [
r1
j f O dx
~5 f 1
I\~
..,
j2n' a
Reject Ho when
.L
1
O
~ (Xi
'"'
l.
known
~)2
f
1
O
I:"
- -
J2n'a
~
2
~
2
- 2u..
.:::JX.
""
nu..
..
~Xi
.' J.
J.
".L
,
2a'2
>k
~
or
to dx
IJ.=~> 0
2&2
~X.
is u"mQpO
a
... ~l_
to dx -
~0
dx
Example I i
dx
>
;?
-151 -
(In k) 20' +
or
nlJ:t. 2
Remark:
Pr[ aJ,pr
X > zl....<l.
J
~)
X is NCO,
H:J..
of
IJ:LJ
=0;"<
= Q1
is independent of
for some family~" then the probability ratio test, *~ is u$m.Pe for
He'
O against~.
Q =: Q
Therefore"
For Hoa
X > zl_..<
<0
and l.J.
>0
uQom~po
,
u~m9P.
It'-
<0
11.:
j,l.
'> 0",
X< zl-'.
J;
differ.
here&
+\
'j'
.'\
Problem 68:
6~t
IJ.. 0 against ~.
There can be no
l.J.
.fn
X< K'
u.m.p, hereg
~4
X> K
--._---
l.J.1
f(x.)
= Q e-Qx
x '>
(1)
Find the u"m.p~ test explicitly (i"e., find the distribution of the
test statistic) c>
(2)
- 152 Remark:
--
flex)
fo(X)
variable~
lOt
lit
.05
Pr(~>~
exactly.)
=1
flex)
when i'O(x)
>k
=0
flex)
when l' (x.)
<k
I:
when
rflex)
(x)
o
=:
~(k).
I
.t.(k-ltQ)
'----
(k)
(k .. 0)
Problem 69g
(1)
> 0(1
If So and Sl are not coincident then there may be no test of size ..<.
<.,( with
eGO,
Xn are.NID
(~, cr2)
If
power
Cy
=1
13(Q) = Pr[rejeot H]
for HO: I-.t. =: 0 against
I1.1
I-.t.
>0
Testg
..
,,(
- ---
@"'o
w.
i).
~ (x, Q) d :\(iI)
-=
w
NV
Let Hobet
I1. $
i~e~1
Let
Th~~
34:
Proog
= Q
%against ~.
If rp~ is of size d.. for the original test" it is mOp$ for this test.
(x) f(x,
Then
'"
=
Thus
andl
is
J[5 ~
iii
Ha-
(x)
f J(x~
lit
d:\{iI}
iI) d:\(Q}}
~ (x) ~ (G) dx
AA-
Q} dx
(x) (x, Q)
dx
dx
~54
ci known
rf' test~
H~:
!f
(;J.
I:
I(. /, ]
Put
rn
0 against Hl :
(../,> O.
~"CO<
O.
for(../,
A(Q) = 0
'3 <0
-1
Q>O
,/
~ (x)"
( :rex.
~) d},(Q)
..
:rex,
0)
so that our problem is back to HO and we have for HO a u.mop" test which
is also a test of the original HO'
One way to get an optimum test in the absense of a u"mop" test is to restrict the
---------class of tests to be considered and look for probabilH,y ratio test for restricted
class.
1<)
..Defo
42~
2/j1
Def. 43g
Unbiasedness
rJ
co<
for all G e ~
Similarity
is a similar test if Eg (
rJ)
Wo
3e Invarianoe
X has density f{x, ~)
G :=, a family of transformations of the sample space of X onto itself\)
eo go a g(x)
=:
cx.
= x + d
change of scale
translation of axes
g is
g(Q)
JL
g(~.f (12)
if X is
lit
OJ,.L,,
N(~, 0'2)
g(~,\1 (12)
X ... d is
lit
~ e wand H.J.I ~
IT we have HOi
oX is N(0 j,J."
2
0 0'2
~+
2
0'2 )
ci
d"
an. w
8: (,.J
i( i)
8:
8: UJ
n... _
~ when
0'2)
Example:
when Q
H08
Q oS
~.
n .. w
g(X) .. oX
>
-----+-----1-/.
Def.44t
eu
Example t
t.
g leaves
l~
n
-"i X.
J.
=:.
lit
~(x)
l~
(n-l) n
(under g
ii ~cXi
ex) ---"':::-_~r-J"I\'
(~~i_Iq.)~\1/2.
(n-l) n )
ii~:'X'
lit
~(Xj-f)2J 2
(n-l.) n
Def* 45:
rjf
Single parameter Q
If
=:
~G
is unbiased, then
=:
Proof:
~ .(~)
~_ (Q)
~
>-t
Hence
for
~ ~O
~_ (x)
(x, Q)
d~
(Q) ..
'aQ
dx
Ill:
G .... Q
O
Assuming that
He:
=:
and henoe
Notes
"Q., 9
.Theorem 35:
5f(x, Q)
11.$
@l (two-sided)
If there exist
(
I:
I:
~,
we can get.
so that
~~
"I
Q
=: ~O
0 elsewhere
(refS
Cramer p.532)
Let be any other unbiased testo
Then
Sfu fa dx ='" =
J'(w: ~"' Q
'='
90
)_ fa
dx =
dx
a=
J'(- ~ I.
dx -
= Qo
Bi~ conditions
unbias~dness
conditJ.ons
- 157 -
)Cf -lI)
f 1 dx
Cf - ) [f1 -
kJ.
~~ f 1 dx ~
f 0 - k2
.
~i
k:J.
1 dx
~ IQ ~ QJdx~O
f O - k2
i =
~0
~O
1.
lit
#}
>,
o=f~ti
~ b which max~mizes
f i dx = c
i = 1,,2, "",9 n;
=a
otherwise
Exampleg
,
!.
~,
known
lJ:Le
~ (Xi~)2
2 -
2a
- 1.58 -
is~
Reject H i f .
..
(derivative evaluated at
~x.
~.
J.
--r
cr
nlJ:L
. -
~ k +
r
n~
-:r-!
e
n~
.......
kJ.
: 0)
2a
,-
where,
+. k 2 x
k'
-""1
=:
2a
n ~2
k'
e
2 =T
a
2
2a
If we set Yl = the left hand side of this inequality" Y2 = the right hand
'1.. ~ 0;
graph~
side~
follo~
~l
\
J
- . - --.+
I
I
I
I
I","
'"
___-==-_......:"'
..,-__-l-
'"
It
,It
b may be +
.
41'
II
Iff
CD
00
0<
specifies thatll
... l.-.,{
rn
i.e."
.--'.--
,r2n 0'
me
20'2
di =1. ..
..<
d (
dPl
'-
orl
d~
rn
f2io;
1. ..
e"
or
,f2na
_rn-
-{2ncr
2ci
..
e
):
00
dX+
rn
$0'
-co
- rn
n(i - lJ.:L)2
n(x - lJ.:L)2
2ci
..
-0
dx
'2=0
n(x _ ~)2
.. - 2(i
lJ.:L
..2
nx
2r:i
=0
dX
..
nx
ill
di
[2]
==0
It
[2 ]
[1]
..
thus become sg
1x
Reject Hii'
>
IJ.r
it is U1>m.p"u.
Problem 70.
Find the
u~m.p.uo
note~
Theorem 36e
~ (Xi
If
test
test for H
O
-
~)2
2
is sufficient for cr
Proof:
define
Y(:) = E
by definition
Thus 8
[. (o!)
'. need
( TJ
be considered.
which is independent of Q
- 161.
-Invariant Tests
Refs:
We have8
Transformations:
parameter.
...
g(x)
density&
has density f
._ X t
[!.,
f(~
Q)
g (Q)]
Gg
Examplee
-g (Q)
$,
Wo
if Q e
""g (f})
.~
if
~ 8
~l
2
Xl' X2, " ... "', In are NlD (.I., (J )
t
=OX
g(x)
:0
+ d
CX 0\+-
g (.I.,
11.~
If
...
> 0,
(C(.l.
+ d, c 2(J2)
:(0
(.I.
is NID
Xl
(J2)
oft-
d, c 2(J2)
~ >0
g(x) ... ex
~,
..
'"
cr2
are
X and S ... ~(X.
X)2
J.
rs
=m(x)
>
- 162 Theorem
37~
The necessary and sufficient condition that the test function (~)
be invariant under G is that it depend only on m<.~).
Proof~ 1)
(2)
Y(m(~) ]
I:
rJ [g(~)]
2)
= (x)
,
a function of m<.~~).
Given [g(~)
(!)
>
(~) == [g t (x2 )
thusg
(~)
:: (x2 )
t
g, call it g , g (x 2 ) == xl
I:'
to prove.
Returning to the IIIStudent D problema
It remains to show that t
... i
JS
is maximal invariant
(l
""I
Setting t :. - "
rs
==.1.
:c.L.-
J'Si
=t II
we can find
_)
-,
Consider..!.. and call this ratio
"a""
t
2
or S = a S
But this is just one of the members of the original family of transformations so
t is maximal invariant.
Hence the problem is reduced to finding the u.mGp. test based on t.
In summary.
Xl' X2.9
H
3
'0
eo.,
/lk=0
Cf )
X,
SW
fb
at
CX, C
:> 0.
-------------------------------------------.
Distribution of Non-Central t{t-distribution under H)a
Refs~
.'C 5
Neyman -Ii- To:tmrska, jASA 1936, pp. 318-326 (tables for the one-sided case)
Welcl?- + Johnson; Biometrika, 1940, PP. 362-389
Resnikoff + Lieberman, "Tables of the Non-Central t-distribution ll , Stanford
University Press, 1957
the- non-central t-variable, is defined bya
z is N(O, 1) .
where I
d.f~
w is X with f
8 is a constant
>
='C
CO,
= rn
f)
i
CT
-rhE
2
(j
when H is true
o
If ~ is true, lJ.
lit
11.
and
Jil (X-/.L:1.)
(]
. is N(O" 1)
(n-1)
2'
-1
'2
Q)
<.Z
<,co
w ~O
't"
't"
= _e +8
rr
Thuss
1 /fU 't"
co
~
o
To get the distribution of the usual t, put
=0
r rr
8)2
!:! _ ~
u 2 e
O.
;1 - \
~( l +1)
du
e-ax xb-1 dx
o
Thus,
ref:
1 +t
Cramer, p.238
du
- 165-
EO
Note.
if &
Since t >t o when the above ratio> k" the probability ratio test thus reduces to:
Reject Ho when t
>
to
Reject Ho when t
This is independent of
Example.
'1.
fl: >
=~ >
~o
is.
t 1-< ("..1)
(~"
Xl" X2,
000"
Xmare NID
Yl" Y2"
000,
EO
against Hlo
el)
2
(J
i~ u
sp
==
~ (Xi
- X)2
m+n...2
+ ~ (Yi - 'Y)2
V\
\;"-0
\l\
_ 166 ..
Consider.
N( a
X'=a.X+b
Xt i
yt : oY + d
yt is N(c
+ b, a 2a2)
~2 + d, c 2a 2 )
when
a~ +b>o ~2 +d
b dJ
requirementthat~
a=-o
Xi' =- aX + b, yt == aY
oft
b, a ')P 0
To be provent
is a maximal invariant statistic.
t =- t
X'= aX + b" yt
=- aY
-11
x. - "
=-
,?
define:
o(x .. i)
now let XI
or~
==
j( t
..
!I
=d
eX
o(x ... Y) =
ef ...
.. ei
il
'if =- of
Xf
T~e
..
=:
l:<o
d ..
d -
+ d
eX ...
iSI
Rejeet II if
Y': aY + b"
"> 0
-161 ~
- 1J.2
I:
d" thus
(!..:-Xcr - d)
s
cr
"'"
JH
m. .
T
!!
cr
1
Ii
crJl:+!
m n
.
Exercise.
.( == 0.05
-cr = 0.8
Find m" n required to have the power
'?
Answer:
non-centrality parameter =p
2(n-l)
=0
-30n
28
27
25
For the power to exceed e90" from the Neyman-Tokarska tables we need:
= 30
d.,f o = 00
p == 2.99 at dGf o
P ., 2.93 at
a)
crl
= 26.7
or 27
Xl' X2,
.
Y
,
Yl~ 2
HO~
0"00,
"Ill""
2
== cr2
1\$
crl
>- cr22
b)
c)
d)
-168 ..
e) Find the u~m~p.u.i. test (ioe., set down conditions to get the rejection
2
2
region as in problem 70) forI HO' cr1 = r:J2 against H~U (J12 rJ cr 2
2
f) Show that the usual test which is to reject if F '> Fl _..<. (m-l, n-l) where:
s 2
1:0
~ (Xi"
x)2
m-l
~ (Y ...
z 2
y
.~
!)
n-l
1=
= 0.05.
~ (Xi ..
~) IJ. known
Reject
Ba
it
~ (Xi ~ 1J.)2
2) IJ. unknown
is sufficient for
/.1-)2
)t K
< IS.
~(X.
or
-
>K2
xl
J._
.,
cr
n-l
--
H:t.
H~.
cr
-It
= f(X
X.
Hence, since (n-l~ s is ,,2 with (n-l) d~f. when H is true, the problem .is
O
(j
exactly that of (1) with the
d~fo
reduced by 1
(i~eo,
-169No~alTests:
Summary of
Xl~ X2'
Ha :
HI:
~ )0
&It
Test:
(~
Reject H if
Classification
Invariant under
(f or H against the transformations
given alternati va)
of the fom
known)
x )zl
O"x
-..<
rm
-
O"x
2 fl
"
I-I )Zl_~ 1m
-
Hi: ~
:f
HO: ~x
= ""0 ( O"~
HI: ""x
> ""0
t ) t 1_..<
H'I"" ""x
I: ""0
It
IX
= i0
3. Ha : 0"2x
HI: 0"2x
2
Hr: O"x
1
unknown )
Xf = cX
I > tl~2
e)
Xf = eX
e arbitrary
("" known )
2
~ (Xi .. ",,) 2
>' 0"0
F 0"0
2 (m)
> X1-..<
--:1
u.m.p.
>K2
4e
HO: 0"
= 0"02
HI:
O"~ )
HI:
O"~ F O"g
og
("" unknown)
~
- 2
(Xi - X)
~ (Xi
2
> X1-..<
(m-l)
>
- j{)2( Kl or 1<2
(for K , K see problem
2
l
70 - use m-l d.f.)
xr = X + a
Xt = X
-I-
-170Alternative
Hypothesis
Test:
Re ject H if
H : lJ.x :: lJ.y
O
x' = X + a
I'=I+a
x' = X + a
I' :: Y + a
x,
a.
'f, s~
m+n-2
x-f
x' = aX
= aY
yl
sp/ ! .
m+ !
n
I X - y I > tl_~(m+n-2)
2
s p(lii-r"ii
J!":T
ix r iy
X, f, s~J
+b
+ b
Xl :: aX + b
It :: aY + b
~, ~,
0';.
X-I
s2
2:. + _l
Ref:
Tables far t I in the one-sided case (-< = 0.05, 0 ..01) have been
given by Aspen in Biometrika, 1949.
-171If m = n
X-I
= -:;:::::::::;;;:
+ s2
ix
I
V
X-I
Empirical results:
based on 1000 samples"
= n = 15
cri = 1"
cr~ = 4 :
significance level
actual rejections
5%
1%
4.9% 1.1%
actual rejections
5%
1%
6.4% 1.8%
cri
vfith n l f n "
f cr~, the U test stays fairly close to ..(
2
in size, whereas the t-test jumps wildly ani a comparison
of the power becomes very difficult.
-172Alternative
Hypothesis
Test:
H'cl=i
O' x
y
Hl :
2
x
(j
>
Reject H if
Classification
Invariant under
(for H against the transformations
g!!~m alternative)
of the form
known)
2
y
XI = aX + b
yl = (a)Y+c
(j
XI = aX .,.. b
yl = (:ta)Y+c
~2
2 )
1-, 1
sx
>F l~
(m,n or n,m)
8"
(lJ.x '
~y
unknown )
u ..m.p.i.
XI = aX + b
Yt
u.m.p.u,.i.
= (ta)Y +c
XI = aX + b
yl = (ta)Y+c
- 1733..
Ha:
~:
Q 8
n-
max f(!I Q)
Q 8 Wa
11.= max f(~ gj
8..0..
For small samples in general this procedure may not give reasonable results.
satisf~
For
Proof:
Refi
Example
of
...
m,l.r~
Walda
tests
2
Xl" X2 " ou" ~ are N(~" 0.)
HaS
=a
~Fa
R.t.8
~ (Xi
2
20
- ,,)2
unknown
-174 ~
~X.
When H is true"
J.
X"
And in general"
n~X2
----~X
i
- ~ (Xi" X)2
An: ~X2
2
n~
(X._X)2
_ _...;;J.
_
2~ (X -X)2
i
(n-I) s2
(n-l ) S 2
-it
nx-.2
.if
.-n2
.
:::
h(~) :::
of Ae
+~n:r
n )'=2s
J n-l
- 175 -
(~)
hf(A) ==-[n-l
h(~)
Therefore
(X 2/ n _1r l / 2 i\..f/n)-l
(~) <
0
~o
h('iI.)
</\o
h(:\) > h
o
ltl >t o
Reject H whens
O
against
lJ,=O
1\:
I-L
> O.
Xl" X2,
Hog
0>01
(j
-::: C
lJ,
lJ,J
(specified)
4.
General Linear
(j
HIg
S :::
== c 1
ci unknown
'>
Co
~.
12..
00
=:
1, n
II;';
20
00
~" N
Hypothesis~
assumptions8
lJ" ==
3.
j=l
a ..
3.J
i = 1.1/ 2"
~,
J
~1"
$.0,
~p
p ~N
rank A ::: (a
ij
) is p
~l"
13 2"
ClU"
i==l
".ei lJ.i
=t
.f, ::.
1" 2"
u.
N..,p
HOtJ
i=l
C1'...
'10.
~,3.
= 0
k -1" 2"
GU',
- 176 ..
Example;
i.e.~
=- 1" 2,
eo .t
ab observations
'"'1." ~.,
0<.
j=l
parameters in all
It:
b -
a-l .
Lemma:
fl,
~ c~. ::
j=l
ProofS
for i
l.J
= 1,
2,
..
"
CI
o~.,
.. 1.7'7 ..
c 2j ::: a 2j ..
~lj
~clc;.
e~cla2'
j
J J
j
J J
..
A~ci,
j
J
~clja2'
j
~Cl,c3'
, =~a3,cl'
j
J J
j
J J
This = 0 i f
Similarly
A:t
=.
..
~a3 .c l
j
'
J
.~ =~a3jC2j
J
Finally we set:
--------
~~-
-1-78 hypothesis~
i=:~
A. ti
~i
Ill:
From the lemma we may assume these form the first N-p+s rows of an orthogonal matrix,
These can be extended to form a complete orthogonal matrix (NxN) -- this is a well i
known matrix algebr~lemma,\l proof is in Cramer or Mann. If we call the complete
'
orthogonal matrix ./ '- , we have
.0.
~l
..
II
"N.p"
III
Pu
.0.
o
"
":w
~-p,
Pl N
'A,ts-"
P Vs orthogonalizE
..
0
C)
fJ
PaN
Psl
III
'l;lN
"
<l>
'l;
p-s, N
E(Y,e)
= 1.=1
~ 4 Di ~
'CI
by the assumptions
=0
.0
if H is true
O
ci
yts are independent normal variables with means as shown and variances
(an orthogonal transformation changes NID variables into other NID variables with the same
varia.nce s) "
].
H:t
where lJ.i
at:
=- 0
I
"
S .;.
=- -
( {2i1 a)N
m.loe. of the IJ. 's in ware found by setting IJ..]. = Y.]. (i=N-p+s+1"
N-p+s
and~2=~
1=1
A =
u."
N)
y. 2 "N
].
8~1
(~
e - N/2
e - N/2
N-p
~y.2
i=l ].
N-p+s
~y2
i=~ i
Qr
N-p+s
y. 2
== Q + ]
a
i=N-p+~].
'A,N
Q
a
CJ;
=(~r
= absolute
minimum (nothing
specified about the hypothes
r = relative minimum
-180 -
Qr - ~a
- 'l:9
1'1
A. -1=
X (s)
N;f
=~
y.
i=1
Therefore
~d.2
N.;fItS
Recall that
is
J.
i~-~l
>
(S.ll
N-p)
J.
x2..distribution
i=1
d. 2 }e
J.
(Q - Q )/s
The ratio of a non-central X2 to a central X2 ia a non-central F ~ thus ~r_,...-a
__
Qa P
s
under H.. has a non-central F-distribution with parameters (a, N-p.t ~ d. 2 ).
--~
i=l J.
7N-
Thus the problem is solved in terms of the yts (the orthogonalized XiS).
The original problem was:
XiS are NID
..
1)
i=l
)..0'
",J.
(~i' a2 )
~i
= 0
2)
i=l
Pki jJ.i = 0
.e = 1" 2"
Q $
N-p
- 181-
HeLoR. test:
defineS
Q;
Q;
~
=:
min
~ (Xi
=:
min
(in .f).)
G (in
w)
~)2
(Xi - lJ.i)2
=:
(Q~)lJ2
~ (Q r )1/2
r
')/2e-N/2'
.
"
a
"" =n'
-
""r
-Nl ~
Q - Q
r,
Qa
Recall that under an orthogonal transformation, sums of squares are preserved.
Thus:
Q is carried into Q
Q is carried into Q
Hence:
and has the same distributions under H (F-distributioj
and under
,Problem
74:
11.
(non-central F).
j.l.
~J
=~
=~,
2,
= 1.. 2,
i
j
~J
0.0,9
a
b
+ ~i + ~. + (~)i'
J
J
all ~.
H
O
2)
= OJ
all (-<13)
~J
=:
- 182-
Tables ing
lit
2~JJi2
s+.
Mann
Kempthorne
s
2
aO' A
==
d.
k=l
i~
YN-p*k =
i==l
P ki Xi
~ ~
==
Q-Q
11,=
--,-
=~y'
k=l
(N
=~]
N-p+k
k=l
i=l
I:
IrA.
Example:
or
X" k =
1J
f,J.
13J
]13. ""
j
-I;
(0<[3) 1J
e"
-I;
1J k
~
(0<[3) ..
i
1J
..<. "" 0
"" 0
X eo. l
under
~,
Qr .. Qa ==
~(~13) ..
E(l.1 ., ) = ~ + -<.1
E(l
C'
) ""
E(l.1.. -
+ 0 + 0
1J
=0
- 183 a
hence
~ ~
20'2",:::
thus
Exercise:
..{.
nb
1.
.,(. 2
i=l 1.
I:
[ 20"2
If a
I:
Randomized blocks;
i .. I" 2"
j = 1" 2"
~..
J.J
~ +
J..
oft
.. e ...
"0.''
b.
X,.
J.J
I:'.~
where
+ "<i + b. + e ..
J
J.J
ij
j
is
N(O, 0'2)
is
N(O, O'b 2 )
-1...1.
rj.
~b,
Given b."
J
If H is true"
X.1..
ib
X.
1..
is
N(~ ' ..
i=l J
II>
2
0' )
""0
0'2)" and
~ ~ {f. - X
1=. 'J=
1.'11....
a
)2
~ (X. _ X )2
'11.
1.=
2
has a X -distribution with a-1 dGf. This conditional (on the b,) distribution does
not involve the b j " therefore the unconditional distribution of J
a
~ ~ (X, - X
i=l j=~
1.
)2 is
x2 (a_l).
..
...
-184 ...
~ ~ (XiJ.
i
- Xi....
X.
-J
2
is X (a-l)(b-l).
)2
He
ft
.>y'a
(ili. - il
~ (X'j~x-.-'"';"'--x-.-It-X-)2-~~(a---~)-(b-"-l)
[b x.1"
is N(IJ.'
+$ ..<.,
1
~c(i2.9
.~~'
.J
1.
-1
(a-l)" (a-l)(b-l)
dof.
~ ~(X1.... XO'~ )2
ri)e
has a non-central
dot' ~
a-l~
00.,
i =~J1 2,
j = 1.~ 2" ..
where the a
u,
a
n
ANOVA Table
~
aSs
error
a-I
a(n.,l)
E(MS)
SaiS.
~ ~(X ....!
)2
~ (X i {'X i )2
1.
CI<ll
+na
unconditional distribution.
j.tb
~ (X. - X
i=l
J. 1t
)2
Qa
a2 itnaa 2
He;
eTa
a leads
...
~g
--:2.
a
+na
a
Xijk
1ll:LJ
a)
b)
I.kij
=:0
,i)
is N(l';.j'
l..l.
-I;
it
b ..
J.J
a. + b ..
J.
ij
J.J
..,
= 1,
= 1.,
= 1.,
a.
2,
b
2, .,
2, 0 , n
k
where b
are N(O, CT 2 )
ij
b
2
where b
are N(O, a )
i
j
a.
J.
fj.~,
are N(O,
CT
2)
= 21
= 2;
1.)
2)
a =
5;
: 0.05,
in (a) against
in (b) against
eT2 +
aa
2~ 2 2
-186
~
MUltiple Comparisonsa
Tests of all contrasts with the general linear hypothesis model following ANOVAI!>
Tukeyts Procedures
observations:
let R =- max
~
se
Xl' X2 ,
i Xi ~
~
'5 i
is an estimate of
eo.,
Xn are NID(O,
- min {Xi}
n
1. ~ i ~n
2 which is independent of Xl' X ,
2
qU,
Xn with
f d.f.
(iQe., f s 2;02 has a X2.distribution with f d.fQ)
e
The distribution of Ria can be found in a general form" and in particular when the
X9s are normal (since the X.la are N(O, 1) -- it is free of any parameters.
:L
The distribution of Ri
RI!a
= !..
s e "'" s e
::0
Studentized Range
numerical integration and percentage points have been tabulated by Hartley and Pearsor
in Biometrika Tables, Table 29Q
Consider the one..way ANOVA
situation~
i
j
2
!i. are NID (I-l-i , ~-)
~ ~ (Xij
2<0
Xi )2
a(n,,",l)~-
J.!i::O ~
= 1"
2,
~ ~,
=- 1 ~ 2~ 0". II n i = n
2
is an independent estimate of a with a(n-l)
or IJ.i .., ~ = 0
Given HO:
Pr
d~f.
~87
Frame of reference is the total experiment" not the individual pairwise tests -- is eli,
for the total experiment and making all possible pairwise tests, the probability of
the Type I error is ~..<" where co< is the level chosen for the Q(a"f) statistic.
i.1. - '1<:
Hence we reject H:
I:'
rn (Xi. - X
{Tm
k)
Q..< (a"f)
l.
1, 2,
1" 2,
with f dof.
~ c.~l.'
i=1
...
i=1
l.
X.l..
0 ....
l.
HO(~).
Pr
(a-I). C
a-l
where 0'
()
[ i=l 0i
(Xl.'
----
- X
>}
2
s 2~ 0i
e. i=l
n.l.
~-<.
a-l.1
F"l
(1
r(a-~+! )
r()r{f)
_ a-ltf
F)
dF
- 1.88 Proof:
Under --0
H_
~c.X
1. ...
1.
1.
0,
=:0
>
~Pr
.......
~c i =0
Y.1. are NID(O, 1)
define:
i=l, 2, ".:1 a
d.
1.
tit
(]
d. Y)2
1(,2
1.
i_
with respect to d
r~ rn:J y.J~di
Y~li=
J
~n.J
multiply
[lJ by dj
since
~ (ltd.
1. 1.
:=
since: i)
~ v:'d.
J. J.
ii) ]d. 2
J.
=- 0
=:0
- J.89 put
":t"
2Yj
nj
= ~
(Y . - i)
J
j=l
Reoall:
Pr
.,
J.=..
c.J. !.J.. ]
t
~C'!i
J.
,
:il:. Pr
>t
~ n. Y.
_ _...J.---.;J.;;;.._
~ni
:=Pr
6e
Y.J.
22
(]
>t
Then
Hence
>a.:rt
Thus
but
Therefore the type I error will be less than or equal to ..< if we reject any H (!!) ifa
O
L~Oi 1 '>
2
Xio
se
2~ci
~-
ni
191 $"
~.Tests
Simple hypothesis
classes (or
pr6babilities~
observations:
l'J.
P2
vI
"
Pk
~P.J
~v j
=1
=n
If H is true
X.
lOt
(
0,
"'Iv. -flP.'/
~. -1L.~ 6~
j=1
l1P
j
--71
as n--7oo"
p. = p.
J
v. - np. o
V ...
np.
=_L.-...:2
X. =- -L-_:.L
.r::-Or:;::'J
p
, np
'I n
$ j
p.
now e
-:L
0
Pj
p.
1. + -J
Pj
to>
p. 0
J
as
Under HI the Yj are asymptotically normal!)
n~oo.
- 192 -
~ Xj 2
under
H:L
is a sum
of squares of Yj +
Pj
k
j=l
k
n~
j=l
Example a
P .. Oc4
p: 005
p.
No .. boys
Pj ~under
under HO)
.16
~48
.36
A=n
=t
I1.)
,,0816 n
For n = 100
The power can be determined from the Tables of the Non..Central x 2 by E. Fix,
University of California Publications in Statistics, Volume 1, No.2, Pp. 15-19.
From the tables, for
00 7
~ ~
~kw
for j
> k'
- 193 -
x2.teat
o2
(v, - np, )
caloulate ~ a J - as usual
is modified as follows:
for j .f k':
nPj
for j
>kit
<.
if v
> np,O
J
if v j
x2
if v j
<
np j
np j
=0
calculate x2 as usual
Examples
8(10)
12(10)
20
12(10)
8(10)
20
Refs:
if v1I
X~
=0
--
0.5
=:
therefore we fail to
reject H
O
20
20
noteg
Hog
> 10 we would
then calculate x2
1
vll
'"
0
probabilities:
2
v12
0
0
k
vlk
.eo
"
v sl
v s2
..
P1I
P12
oQ
"
v sk
Plk
"l)
PSI
P.s2
.,,.
Psk
= 1, 2,
2:1
= 1,
...0",
~
j=l.
Pij
Vi'
J
=:
=:
f(~)
s
k
- 194 -
~
[P;;
j
~J
HO~
Pij
X' j
~
1:0
for i
= PijO(~)
Theorem 39.
1\
IrQ is estimated by
then as n
~ 00,
r: .
1\
""\2
LV i; - ni.Pij (~) I
-~
7t
n-lp, . (9)
~J -
-=-
.l.
If H1 is true8
Pij == Pij
=:
Pij
c. .
-8-
and
ni
=:
--
~n
l/ n i
i
2
then X has a non-central X2..distribution with d.. f. as before"
l
and non-centrality parameter .S [ I B(BfBr
where Oi
- ], x ks
(..JC"JPi)
Cramer:
Mitra:
J~
:: _~J
r=-o- ~
Pij
Ref:
B~
i = 1" 2"
j # 1" 2"
.e == 1" 2"
0'." ks
.. /) 0"
" "t
..
x2 section
December 1958 Annals of 11ath Stat
Phq,D. Dissertation, UNC, Institute of Statistics I".fimeograph Series
No,,142.
- 19, -
Jn
11 JPij X1j ~ 0
rel>i;rictionB that
~ 1,
for all 1
2, .... B
1J
dIn
"$Q
I=
dInQL
-a
==
+
Q
(~~~2In L j
oQ
~ =9
) (9 -
9 )
0
Recall:
1n L =- KI
01;
~~
i
...t-~
'0 ili
h =- 1" 2
Hence the estimation of Q (single component) means one additional linear restriction
on the v. . or on the X. j (since the two are linearly related) It
J.J
J.
~~
ij
Xij2
into
~~
y ..
ij
with s(k-l) - 1
1:1
J.J
Probabilities:
1)
11.1
2)
Pu =- P21 =
~:
P12
=P22 1 - Q
P11
= Q + o/[ll
Pl2 = 1 - 9 -
.P21
= Q - olin
P22 ==
1. -
ct F
c/
rn
A =t,2'
take PI
= P2 '21
(sampling fraction)
::0
~ (-r;--Q-
BI
BtB
- ~96 ..
::0
-1
(BIBf
=:
-Q(l1-
=-
1 - 9
Q(l - 9)
~
.. Q
- 9(l-~
2
1 -
-Q(1-9)
... Q(l-@)
1
B(BIBr Bi =.
1 - Q
-r-
(~
5'
I 0
T
-c
J2(1-9)
2"
1 .. 9
- Q(l- Ql
2
-9(1-Q)
-2
--r
- 9(1-9)
2
~--
- 9(1-9)
2
-rQ
~-
.. Q(l-e)
-61
Q)
-2
-c-
-c
~2(1..9)
::0
;..
c2
= Q(i-Qj
of;
P2I
=---r2c
PII" P2I
= Fn
PI2 + P22
I-Q~--~-
n (
=t 4
Pll .. P2I )2
- 197 -
A=
In the Deoember 1958 Annals of Math Stat Mitra gives the following formula for A in
the 2 x k contingency table&
probabilitiesg
~Qj
J = Q + c1.J./
pJ." "
rn
=1
~C1j = 0
~C2j = 0
2
(c lj - C2j~
9j
For the above example:
Pl
= P2:
"2'
sup ~
e ':f'(.1..)
(Q)
.I..
be
<p (..<).
- 198 -
it is of size
ii)
20
I'1inimax~:
L [D(X)"
Example:
X is N(Il" 1)
D(!) == 1
D(!) ... 0
means accept
HO~
Il'" 0
Having set up a loss function, we then have a risk function defined as:
rD(Q)
~ E(L):
.5
00
(:>(!:), Q]f(",. g)
Q"
thus:
SUPg reg)
or expressed another way we determme3
ref:
~ rD(Q)
for all t;
rD'(~)
<::
for some Q
rD(Q)
---------
MISCELLANEOUS
A.
Confidenq, Re gions:
B(w)
Let
Def.
ta.:
=:
A(~
If Pr[A(!>
Theorem 40:
.rL
::> e J)
~
1 -.,(
...
then A(X)
~.
exists for H : e
Let R(X.
-
e)
90
_9
_
=:
for every e
0 0 0
confidence coefficient 1 -
=:
e of
size 1
J ~.
eo is rejected, e.g.,
R(X,e),
lIl1f111,
IllIflfn
--x
[lJ
U X J R(2,
./
e)
J R(e)]
x.
=
1 - Pre! e R(e)
J >
1 - ~
q.e.d.
::>
Ie'
e' r/ e.
note: Kendall uses this same definition with "uniformly most powerful"
replaced with lIuniformly most selective. 1I Neyman replaces "u.m.p."
with "shortes t II
Def. 42,:
,~_==-.
a~, a~, p)
= correlation
wherejO
Problem:
i = 1, 2, ... , n
= O.
= ~fi~
Zi= ..<Xi -Y
i is N(O "~2a2
x - 2..<na
r x a+
y 0'2)
y
2. (Zi
- 2)2 and
Z are
L (Zi-Z) 2
n-l-
=
222
=..<
sX -
2~s
:xy
+ By
therefore:
4):~ =~~i1/2'
x
Pr
xy
d.!.
-.... 1- -1
x - 2..<s xy +
l-..
2
,,~.
(n-I)
=1
- e
- 201
000
Among the various possible solutions of this quadratic equation we can have the
following situations:
~,l~/,,(
2
A.
If nX! .. t s~ > 0:
~"<2
It can be shown that 2 roots always exist and this the confidence interval is:
...
For the above inequality to hold, X must be significantly away from the origin;
i.e. , we must
rejectHo:~
=0
i\lx.L>t l
x
.. .!
2
One could, if desired, change t (and thus e) to insure that this inequality
always held and thus that a ureal" confidence interval exists.
Note:
B.
If
nX2
- t
s2
x
<0
~,
If the roots are real, then the "confidence interval" is (- 00, ~), ("<2' + 00),
i.e., we accept "in the tails."
......
2)2
2\ -2
2 2
2
If 4(nXY t Sxy
< 4(nX....2 .. t 2sx'
(nY - t Sy), i.e., b
C.
< 4ac,
then
the roots are complex and we have a confidence interval (.. 00, + 00).
[Thus we can say that .. co
D.
If we get
+ co with confidence 1 ..
Q(..<)
----+-----'-"..<)
Q(..(
' \
Refs:
eJ.
s. 0 <:
ill:..2
(..( s
1..<1 - Y I
2
x
.. 2..<8
2 l7~2
+ S )
xy
Y
<t
- 202 -
LIST OF THEOREMS
Theorem
10
G.
4.
..
22
22
23
8. Tchebycheff's Inequality. .. .. ..
24
9.
25
7e
10.
...
...
..
28
29
12.
30
..
~.'
.....
33
II
"
..
"
< 00 )
<l\
35
41
41
"
44
46
51
61
.. .
33
65
70
- 203 -
Theorem
Page
24~
25.
76
81
26c
88
270
28.
290
109
300
116
o.
124
32~
*' 140
33~
G.
92
96
(>
149
153
156
"
""
162
187
194
".
199
... 204 -
LIST OF DEFINITIONS
Definition
10 Notation of sets
2. Families of sets
Oc
. . . . . .. . . .
. . .. ..
_'.
-0
.,
40
50
Probability measure
6"
30 Borel Sets
..,
.... .. ..
0
7. Distribution function
8. Density function
<3
9. Random variable
'.
..
..
..
II
_e
"'0.
..
..
-0
O'
.e
cD.OO
10
0.0.
10
II
O.~
16
-0
..
...
GIi
(0)
10.
Conditional probability
ll.
000,
14~
Expectation of g(X)
1$"
Moments
160
Characteristic function
17~
180
9-~e
()
(l
"
17
21
It!
..
O.OeJ
2$
..
28
hI
"
"
"
"
..
"
eo.
e.
II
"
..
22
Risk function
..
21 0
I~nimum
22.
230
Bayes estimates
24.
Minimax estimates
II
..
"
<)
..
ill
..
..
..
"
"
(m.v~u.eo)
.. .. ..
..
" "
..
"
"
"
..
I>
"
$5
56
57
58
60
- 205 Page
Definition
Constant risk estimates
26.
Consistent estimates
OIl
.Oll.,~
II
eo
60
o.
60
60
32 0
330
Test
0."0.,000
ooo.o
"
"
e"
...
40.
Sample quantile
"
o.
64
oa
..
86
92
113
"o.o
'
113
113
.'
o.o
eCJ:o.
CI~.o
. . .,
"
..
l).
113
113
."'.e.o.o.o o
114
tIIQ.e
115
.)
116
1). 154
154
155
64
"
oa.<t;>.o
"
. . ..
61
30.
Likelihood funotion
290
~.
oo
148
42.
Unbiased tests
43.
44.a
Invariant tests
45..
.....
155
46.
470
Confidence region ..
48 0
490
80
..
"
0'
oo
.,oeo
ee
161
..
..
..
..
199
.. .. .. .. .. .. 199
..
0$.0
..
..
...
200
- 206 -
LIST OF PROBLEMS
Problem
o.
lim
Pr,( 8n ) " " " ..
n-?,oo
3.
40
5~
Transformations involving uniform distributions " " " " " " "
1.
Pr(Sum of sets)
Sum Presets)
=0
"
"
""
13
19
21
"
(l
. ., " "
"
<>
." .
. ." .. , .
"
21
24
10.
Prove the normal density function integrates to 1 " " " " " " " " "
28
ll.
"
(II
28
eo.
.,
"
32
"
..
32
33
Ii
10
35
36
38
It
"
38
41
41
coo
16.
"
"
20.
Show E [g(X)
210
22.
44
44
23.
Prove theorem 19
E,x [g(x)ly]
Ci
oo.o.e~o:o
oe
,,~
"
46
48
So
- 207 -
Problem
= cos
ta
eo'
e.eOc
o.
\)
.,
x is a bol.u.e of ~
o \)
<>
..0
~.
50
50
57
57
59
60
..
60
ItI
co
Cla
63
69
0,
"
69
c:
70
74
E>
.,
.,.
74
\)
75
78
Crarr~r-Rao
J.o~18r
~A
<>
lower bound
80
42. Application of the C-R lower bound to the negative binomial ... 84
43. Derivation of sufficient statistics
44.
4)
87
89
<>
. ..
95
..
98
Minimum and constant risk estimates of cr2 (normal) "
98
<>
= p.~1:'.)
J
101
107
- 208 Problem
2
M.l.e$ and minimum modified X estimates of Pi
=e
-~iA
fl
It
"
f}
108
,2. Comparison of variance of minimax and m.l. estimators for Binomial p 110
,3.
,4"
,,~
,6.
1.1.
..
1) variables
. ..
..
tI
"
..
.....
:)
Power of the
n-n
..
fl
i>
..
. . . . . . . ..
"
CJ
..
e 0 0 ..
..
"
..
.......
"
..
..
..
..
..
..
..
.'e"~..O"
"
..
..
'. 112
.. 114
. 115
117
. .. 120
"
II
II
123
124,6
. . . .. 128
Gil
..
, "
. 129
..
134
:)
..
134
138
0""..".. 0
I)
..
"
..
"
'"
"
Kruskal~Wallis H test
138
14,
70.
."
..
..
ct
Co
lJ
. .. . ..
.. "
to
73.
~. 75e
f)
. .. 151
152
160
"
..
167
..
. .. . " . . . 175
" . 17,
. . . . . . . .. 185
147
~pothesis
"