You are on page 1of 8

LECTURE 1

Research in Economics :The Econometric Approach

I.MEANING

At the cost of over simplification we may say that in economics


research means construction of an econometric model of the
following type:

Y=a +bX+ u………………. (1)

Here Y is the dependent variable.


X is the explanatory variable. (There can be more than one
explanatory variable).
u is a disturbance term. It is a random variable .
Equation (1) is the regression model. (It can be nonlinear also).
II.OBJECTIVES

This econometric model is used for three purposes:


1) Estimating the effect of X on Y.
2) Testing hypothesis /ses.
3) Forecasting.
One need not pursue all objectives for the research.

III. THE RESEARCH PROBLEM OF ECONOMICS

We must understand the core of the research problem a


researcher in Economics faces. For illustrating this problem, one
can for convenience distinguish between Experimental Method
(EM) and Correlation method (CM).In Economics the researcher
is forced to use the CM. Yet, she wants to draw valid
conclusions like EM.
Let us consider an example. Let there be two groups of land.
In one group irrigation facilities (called treatment in EM) are
provided. The other group does not get the treatment .The first
group is called experimental group. The second group is
called control group. Let us call yield of rice , per unit of
land,say acre, Y. There is Y for experimental group .We call
this YE. Similarly for control group we write YC . The
averages are called YEA and YCA. .The question is: can we
conclude, from the observation that YCA. is smaller than YEA ,
that irrigation facilities improve yield? Unfortunately, we
cannot. The experimental group may turn out to have a higher
proportion of skilled cultivators compared to the control
group .Similarly there can be many other factors, many of
which we may not have even thought about.
Is there a solution to this problem?
There is. It is called randomization. A truly revolutionary
idea! Ronald. A. Fisher contributed significantly in
developing the statistical theory (Design of Experiments,
1935).Randomization involves two aspects. One is selecting
the respondents or units of study or ‘subjects’ in such a
manner that each unit has equal probability of getting
selected. The second stage is assigning the units into control
group and experimental group in a manner where each unit
has equal probability of getting assigned in any one of the
groups. Now if one finds a difference in the two groups then
the reason must the treatment alone and no other factor .This
is the logic behind randomization. The conclusions about
effect of treatment on the outcome are valid in this situation.

Let us try to analyse the implications of randomization using the


regression function. Let us write again:
Yi=a +bXi + u………………. (1)
We see that Y has three components .one is ‘a’. We will try to
see what it represents .The second is the treatment effect bX.
The third is ‘u’ .u is a random variable. This is the consequence
of selecting Y in a random manner with each Y having the same
probability of getting selected. We think that u is that
component of Y which varies randomly because of the influence
of the ‘other factors’. Let us push he analysis further .Let X take
two values, namely, 0 and 1. X is 1 when treatment is given and
0 when it is not given .For Y in the experimental group X is
1.For control group it is 0.If we take only control group then we
can write
Yi=a + u …………(2)

Now we can also add the Ys.


Thus we have
ΣYi=n a + Σu
Her n is the number of subjects studied If Σu=0 then, we have
ΣYi=n a
Or, ΣYi /n =a= YCA
Or, average Y of control group is ‘a’, the intercept term .
Moreover, u=Yi- YCA. .So Σu=0.Let us turn to the experimental
group. A little thinking will show that a +bXi= YEA. As a result ,
bXi is the treatment effect, YEA - YCA .Therefore we can say that
‘b’, the slope coefficient , is the treatment effect coefficient .If
b=0 then there is no effect of the treatment .
We see that the regression function states that
Observed Y= YCA+ Treatment Effect +Random Effect (of the
‘other factors’).
What has randomization achieved so far? It has helped us to
decompose Y into a systematic component and a random
component. It ha also helped us to see that the sum total of
random effect is zero only when we are speaking in terms of
averages .This leads us to a very important principle of
estimation , namely . the principle of least squares.
IV.ESTIMATION: THE PRINCIPLE OF LEAST SQUARES
Let us call the difference between observed Y and Estimated Y
an error, ‘e’ Thus
ei = Yi-Yc. ……………………… (3)
Here , Yi is observed Y and Yc is estimated Y. We square e and
add to get
Σe2= Σ(Yi-Yc.)2 We want minimize these sum of squares by
choosing a Yc..
The first order condition is that
2 Σ(Yi-Yc.) =0
or , Σ(Yi-Yc.) =0……………….. (4)
or, ΣY-n Yc=0
or, ΣY/n= Yc
or, Yc is the mean or average .
We minimize the sum of squares when we estimate Y as an
average.
We want Σ (Yi-Yc.) = Σ(ei.) =0.But note another fact.
If from the regression model we want to measure impact of X on
Y, then randomization ensures that the other factors are captured
by u. It is also required that Σu=0.By the principle of analogy,
this implies Σe=0 for the observed sample estimation .This is
fulfilled by the principle of least squares.

V. FROM ‘EM’ TO ‘CM’

The economics researcher rarely has the luxury of using data


generated by random experiments. Experiments can be immoral,
costly and time consuming. Yet the attempt is to get valid results
of the impact of X on Y like EM. Analysis based on non –
experimental method can study co-variation and correlation. But
correlation analysis does not go as far as EM in analyzing
causality. To get further the regression model has to fulfill
certain conditions which are fulfilled by experimental method
.These assumptions are precisely the assumptions of Classical
Linear Regression Model (CLRM).

You might also like