Professional Documents
Culture Documents
1 Function of Statistics
Statistics is used for various purposes. It is used to simplify mass data and to make
comparisons easier. It is also used to bring out trends and tendencies in the data
as well as the hidden relations between variables. All this helps to make decision
making much easier. Let us look at each function of Statistics in detail.
The use of statistical concepts helps in simplification of complex data. Using statistical
concepts, the managers can make decisions more easily. The statistical methods
help in reducing the complexity of the data and consequently in the understanding
of any huge mass of data.
Without using statistical methods and concepts, collection of data and comparison
cannot be done easily. Statistics helps us to compare data collected from different
sources. Grand totals, measures of central tendency, measures of dispersion,
graphs and diagrams, coefficient of correlation all provide ample scopes for
comparison.
Hence, visual representation of numerical data helps you to compare the data
with less effort and can make effective decisions. The graphical curve represented in
figure 1.7 and figure 1.8 shows the profits of CBA Company and ZYX Company
respectively, for ten years from 1998 to 2008. The profits are plotted on the Y-Axis and
the timeline in years on X-Axis. From the graphs, we can compare the profits of two
companies and derive to a conclusion that profits of CBA Company in the year 2008 are
higher than that of ZYX Company. The graphical curve in case of figure 1.7 shows that
the profits for CBA Company are increasing, whereas the profits curve in figure 1.8
is constant for ZYX Company from middle of the decade (1998-2008).
After data is collected, it is easy to analyse the trend and tendencies in the data by using
the various concepts of Statistics.
Statistical analysis helps in drawing inferences on data. Statistical analysis brings out the
hidden relations between variables.
5. Decision making power becomes easier
With the proper application of Statistics and statistical software packages on the
collected data, managers can take effective decisions, which can increase the profits
in a business.
Telephone
Online surveys
respondents are interviewed in person, in their homes (or at the front door)
very high cost
suitable when graphic representations, smells, or demonstrations are involved
often suitable for long surveys (but some respondents object to allowing
strangers into their home for extended periods)
suitable for locations where telephone or mail are not developed
skilled interviewers can persuade respondents to cooperate, improving response
rates
potential for interviewer bias
shoppers at malls are intercepted - they are either interviewed on the spot, taken
to a room and interviewed, or taken to a room and given a self-administered
questionnaire
socially acceptable - people feel that a mall is a more appropriate place to do
research than their home
potential for interviewer bias
fast
easy to manipulate by completing multiple times to skew results
Joint probability is the probability of two events in conjunction. That is, it is the
probability of both events together. The joint probability of A and B is written
or
Marginal probability is then the unconditional probability P(A) of the event A; that is,
the probability of A, regardless of whether event B did or did not occur. If B can be
thought of as the event of a random variable X having a given outcome, the marginal
probability of A can be obtained by summing (or integrating, more generally) the joint
probabilities over all outcomes for X. For example, if there are two possible outcomes
for X with corresponding events B and B', this means that . This is
called marginalization.
In these definitions, note that there need not be a causal or temporal relation between
A and B. A may precede B or vice versa or they may happen at the same time. A may
cause B or vice versa or they may have no causal relation at all. Notice, however, that
causal and temporal relations are informal notions, not belonging to the probabilistic
framework. They may apply in some examples, depending on the interpretation given to
events.
Introduction
Consider the simple scenario of rolling two fair six-sided dice, labelled die 1 and die 2.
Define the following three events (not assumed to occur simultaneously):
A: Die 1 lands on 3.
B: Die 2 lands on 1.
The prior probability of each event describes how likely the outcome is before the dice
are rolled, without any knowledge of the roll's outcome. For example, die 1 is equally
likely to fall on each of its 6 sides, so P(A) = 1/6. Similarly P(B) = 1/6. Likewise, of the
6 × 6 = 36 possible ways that a pair of dice can land, just 5 result in a sum of 8 (namely 2
and 6, 3 and 5, 4 and 4, 5 and 3, and 6 and 2), so P(C) = 5/36.
Some of these events can both occur at the same time; for example events A and C can
happen at the same time, in the case where die 1 lands on 3 and die 2 lands on 5. This is
the only one of the 36 outcomes where both A and C occur, so its probability is 1/36.
The probability of both A and C occurring is called the joint probability of A and C and is
written , so . On the other hand, if die 2 lands on 1, the
dice cannot sum to 8, so .
Now suppose we roll the dice and cover up die 2, so we can only see die 1, and observe
that die 1 landed on 3. Given this partial information, the probability that the dice sum
to 8 is no longer 5/36; instead it is 1/6, since die 2 must land on 5 to achieve this result.
This is called the conditional probability, because it is the probability of C under the
condition that A is observed, and is written P(C | A), which is read "the probability of C
given A." Similarly, P(C | B) = 0, since if we observe die 2 landed on 1, we already know
the dice can't sum to 8, regardless of what the other die landed on.
On the other hand, if we roll the dice and cover up die 2, and observe die 1, this has no
impact on the probability of event B, which only depends on die 2. We say events A and
B are statistically independent or just independent and in this case
In other words, the probability of B occurring after observing that die 1 landed on 3 is
the same as before we observed die 1.
Definition
Given a probability space (Ω, F, P) and two events A, B ∈ F with P(B) > 0, the conditional
probability of A given B is defined by
For example, if X and Y are non-degenerate and jointly continuous random variables
with density ƒX,Y(x, y) then, if B has positive measure,
The case where B has zero measure can only be dealt with directly in the case that
B={y0}, representing a single point, in which case
If A has measure zero then the conditional probability is zero. An indication of why the
more general case of zero measure cannot be dealt with in a similar way can be seen by
noting that that the limit, as all δyi approach zero, of
depends on their relationship as they approach zero. See conditional expectation for
more information.