Professional Documents
Culture Documents
Definitions
Suppose that X is a real-valued random variable. The (cumulative) distribution function of X is
the function F given by
F(x)=P(Xx),xR
This function is important because it makes sense for any type of random variable, regardless of
whether the distribution is discrete, continuous, or even mixed, and because it completely
determines the distribution of X. In the picture below, the light shading is intended to represent a
continuous distribution of probability, while the darker dots represents points of positive
probability; F(x) is the total probability mass to the left of (and including) x.
F(x+)=limtxF(t),F(x)=limtxF(t),F()=limtF(t),F()=limtF(t)
Basic Properties
The properties in the following exercise completely characterize distribution functions.
Fix xR. Let x1>x2> be a decreasing sequence with xnx as n. The intervals
(,xn] are decreasing in n and have intersection (,x]. The result now follows from the
continuity theorem for decreasing events.
Fix xR. Let x1<x2< be an increasing sequence with xnx as n. The intervals
(,xn] are increasing in n and have union (,x). The result now follows from the
continuity theorem for increasing events.
F()=0.
Proof:
F()=1.
Proof:
The following exercise shows how the distribution function can be used to compute the
probability that X is in an interval. Recall that a probability distribution on R is completely
determined by the probabilities of intervals; thus, the distribution function determines the
distribution of X.
1. P(X=a)=F(a)F(a )
2. P(a<Xb)=F(b)F(a)
3. P(a<X<b)=F(b )F(a)
4. P(aXb)=F(b)F(a )
5. P(aX<b)=F(b )F(a )
Proof:
These results follow from Theorems 1-5 and the difference rule: P(BA)=P(B)P(A) if
1. F(x)=tS,txf(t) for xR
2. f(x)=F(x)F(x) for xS
Proof:
Part (a) follows from the definitions and the additivity of probability. Recall that
F(x)=P(Xx) and f(x)=P(X=x) for xR. Part (b) is a restatement of Theorem 6 (a).
Thus, F is a step function with jumps at the points in S; the size of the jump at x is f(x).
There is an analogous result for a continuous distribution with a probability density function.
Suppose that X has a continuous distribution on R with probability density function f (which we
will assume is piecewise continuous) and distribution function F.
1. F(x)=xf(t)dt for xR.
2. f(x)=F(x) if f is continuous at x.
Proof:
Part (a) follows from the definition of a PDF for a continuous distribution. Part (b) follows from
(a) and a theorem from calculus.
The result in the last exercise is the basic probabilistic version of the fundamental theorem of
calculus. For mixed distributions, we have a combination of the results in the last two exercises.
Suppose that X has a mixed distribution, with discrete part on a countable subset SR, and
continuous part on R. Let g denote the partial probability density function of the discrete part, h
the partial probability density function of the continuous part, and F the distribution function.
1. F(x)=tS,txg(t)+xh(t)dt for xR
2. g(x)=F(x)F(x) for xS
3. h(x)=F(x) if xS and h is continuous at x
Naturally, the distribution function can be defined relative to any of the conditional distributions
we have discussed. No new concepts are involved, and all of the results above hold.
Suppose that X has a continuous distribution on R with probability density function f that is
symmetric about a point a. That is, f(a+t)=f(at) for tR. The distribution function F
satisfies F(at)=1F(a+t) for tR.
Proof:
Reliability
Suppose again that X is a random variable with distribution function F. A function that clearly
gives the same information as F is the right tail distribution function:
G(x)=1F(x)=P(X>x),xR
Give the mathematical properties of a right tail distribution function, analogous to the properties
in Theorems 1-5.
Suppose that T is a random variable with a continuous distribution on [0,) with probability
density function f. If we interpret T as the lifetime of a device, then the right tail distribution
function G is called the reliability function: G(t)=P(T>t) is the probability that the device
lasts at least t time units. The function h defined below is called the failure rate function:
h(t)=f(t)G(t),t0
Note that h(t)dtP(t<T<t+dtT>t) if dt is small.
Proof:
P(t<T<t+dtT>t)=P(t<T<t+dt)P(T>t)f(t)dtG(t)=h(t)dt
Thus, h(t)dt is the approximate probability that the device will fail in next dt time units, given
survival up to time t. Moreover, the failure rate function completely determines the distribution
of T.
The reliability function can be obtained from the failure rate function.
G(t)=exp(t0h(s)ds),t0
Proof:
t0h(s)ds=t0f(s)G(s)ds=t0G(s)G(s)ds=ln[G(t)]
1. h(t)0 for t0
2. 0h(t)dt=
Proof:
Part (a) follows from the definition. Part (b) follows from Theorem 14 and the fact that G(t)0
as t.
Conversely, suppose that h is a piecewise continuous function on [0,) that satisfies the
conditions in Exercise 15. The formula in Exercise 14 defines a reliability function for a
continuous distribution on [0,)
Proof:
The function G is continuous, decreasing, and satisfies G(0)=1 and G(t)0 as t. Hence
F defined by
F(x,y)=P(Xx,Yy),(x,y)R2
In the graph above, the light shading is intended to suggest a continuous distribution of
probability, while the darker dots represent points of positive probability. Thus, F(x,y) is the
total probability mass below and to the left (that is, southwest) of the point (x,y). As in the
single variable case, the distribution function of (X,Y) completely determines the distribution of
(X,Y).
Suppose that a, b, c, d are real numbers with a<b and c<d. Then
P(a<Xb,c<Yd)=F(b,d)F(a,d)F(b,c)+F(a,c)
Proof:
F(a,d)+F(b,c)+P(a<Xb,c<Yd)F(a,c)=F(b,d)
In the setting of the previous exercise, give the generalizations of the results in Exercise 6. Than
is, give the appropriate formula on the right for all possible combinations of weak and strong
inequalities on the left.
The joint distribution function determines the individual (marginal) distribution functions.
Let F denote the distribution function of (X,Y), and let G and H denote the distribution
functions of X and Y, respectively.
1. G(x)=F(x,)
2. H(y)=F(,y)
Proof:
These results follow from the continuity theorem for increasing events. For example, in (a)
P(Xx)=P(Xx,Y)=limyP(Xx,Yy)=limyF(x,y)
In the context of the last exercise, X and Y are independent if and only if
F(x,y)=G(x)H(y),(x,y)R2
Proof:
(x,y)R2. The converse follows from the fact that a probability distribution on R2 is
completely determined by its values on sets of the form (,x](,y] for (x,y)R2. Read
the section on existence and uniqueness of positive measures for more details.
All of the results of this subsection generalize in a straightforward way to n-dimensional random
vectors.
X with distribution function F. The empirical distribution function, based on the data, is defined
by
Fn(x)=1n#{i{1,2,,n}:xix}=1ni=1n1(xix),xR
Thus, Fn gives the proportion of values in the data set that are less than or equal to x. The
function Fn is an approximation to F, based on the given data set. This concept is explored in
more detail in the section on the sample mean in the chapter on random samples.
http://www.math.uah.edu/stat/dist/CDF.html
Distribution Functions
Consider a real-valued random variable X. The properties of X are described by its
probability
function PX , which gives the probability that X A for any set A R. However, it is
also possible to specify the distribution of a random variable by considering Pr(X A) for a
limited class of sets A; this approach has the advantage that the function giving such
probabilities
may be easier to use in computations. For instance, consider sets of the form (, x],
for x R, so that PX {(, x]} gives Pr(X x). The distribution function of the distribution
of X or, simply, the distribution function of X, is the function F FX : R [0, 1]
given by
F(x) = Pr(X x), < x < .
Theorem 1.2. A distribution function F of a distribution onRhas the following properties:
(DF1) limx F(x) = 1; limx F(x) = 0
(DF2) If x1 < x2 then F(x1) F(x2)
(DF3) limh0+ F(x + h) = F(x)
(DF4) limh0+ F(x h) F(x) = F(x) Pr(X = x) = Pr(X < x).
Corollary 1.1. Let X denote a real-valued random variable with distribution function F.
Then, for x1 < x2,
(i) Pr(x1 < X x2) = F(x2) F(x1)
http://www.google.com.bd/url?sa=t&rct=j&q=&esrc=s&source=web&cd=14&cad=rja&ved=0CDcQFjADOAo&url=http%3A%2F%2Fmath.arizona.edu
%2F~jwatkins%2Fg-randomvariables.pdf&ei=w4kgUq_DIMGVrgedhIGwCA&usg=AFQjCNHOsAQu_Nzd8zxcCoBUrZx1LaYBcw&bvm=bv.51495398,d.bmk