Professional Documents
Culture Documents
Dear all what is difference between ARDL and ARDL bound testing technique?
ARDL is the estimation and ARDL Bounds Testing is like Unit Roots+JJ to verify co-integration exists.
i have taken log returns (LN) of my stock prices, but the distribution of log returns is not normal.
kurtosis is as high as 3000. what shall i do?
Atiq Rehman non-normality is not a big issue and there is no need to force data to normality, however,
such a high kurtosis indicates outlier (s). These should be discarded
Hello! I have an urgent question to You. If I used lag 1 for Johansen coinyegration test, so does that mean
that I have to use lag 1 in VECM too? Or I can use any another lag for VECM?
when we can employ johansen co-integration?
Sayed Hossain You can run Johansen test when variables are integrated of same order, that is I(1) or I(2)
etc.
Frees, E. W. (1995). Assessing cross-sectional correlation in panel data. Journal of Econometrics, 69(2),
393-414.
Wooldridge, J. M. (2015). Introductory econometrics: A modern approach. Nelson Education.
What to do if we have multicollinearity problem?Professor Moulana Naykrasyvishyy
Cholovik commented as such > The best remedy for multicollinearity is do nothing instead of torturing
model with transformations or with poor ridge estimators or with deleting variables. Econometricians call
the multicollinearity 'GODs will' thts beyond the boundary of mankind.
Taken from Blanchard, O. J. (1987). [Vector Autoregressions and Reality]: Comment. Journal of Business
& Economic Statistics, 5(4), 449-451.
there are many independent variables which effect dependent variable, but we select some
variables not all can any one tell why
Including too many explanatory variables has a cost. When you add more variables, you lose your
degrees of freedom (number of observations). Fewer degrees of freedom make it difficult for you to get
statistically significant coefficients. That is why many researchers omit some variables and only include
variables that are likely the most important in determining the dependent variable.
How to remove serial correlation
Suborno Aditya There are several approaches to remove serial correlation. Alternative approaches
include using lagged variables of the dependent variable however literatures suggest its suppressive
effect on overall outcome. Other approach is using first difference of variables but by doing this you lose
dynamism and critical information from your data and hence you may end up with wrong results.
Hence, it is better to use appropriate methods and techniques to manage serial correlation rather than
removing it from the data while estimating panel regression. If it is AR of order 1, Panel regression with
Cluster option (xtreg, cluster) should be alright. Or you can estimate using XTREGAR instead of XTREG
command that by default manage serial correlation. In that case no need to use cluster option any more.
However, if serial correlation oforder 1 is present alongside heteroskedasticity or/and cross section
dependence, using FGLS or PCSE is better. In most literatures, authors do no check for higher order or
possibly moving average serial correlation. Thus, managing for only AR1 where your data actually has
higher order or moving average autocorrelation, will not reveal efficient or accurate result. However, many
authors overlook it and simply estimates equations managing AR1 and gets published. The challenge is
determining whether higher order or MA autocorrelation exists in your data after you have already
identified autocorrelation of AR1. If it can be identified, there are very specific techniques to manage that
in
panel
estimation.
How
to
Remove
serial
and
heteroskedasticity problem
from
panel
data?
first of all we can cheak serial and HSK problem with the help of these two stata commands
xtserial
for
serial
correlation
and
xtgls,
panel(hetero)
for
heteroscedasticity.
Few researchers says that there is no need of auto-correlation test in case of Panel data as we just need
this
test
in
case
of
Time
serious
data.
Here
I
am
confuse.
Yes, if the timespan is not too big compared to n, evidence it will not produce biasedness in
estimates.suupose
according
to
wolrigde
What are the dimensions of your N and T (roughly?) Many of the routines that claim to correct for serial
correlation and/or heteroskedasticity are only guaranteed to work (in the sense of eliminating the
problems) when T is fairly large. If N is large and T is not very large, the "cluster" option after FE -- or, for
that matter, RE -- is attractive. So xtreg y x1 ... xk, fe cluster(csid) where csid is the cross section
identifier. The resulting standard errors are completely robust to any kind of serial correlation and/or
heteroskedasticity. The other approaches assume parametric forms and, like I said, typically rely on large
T approximations. Jeff Jeffrey M. Wooldridge University Distinguished Professor Department of
Economics Michigan State University 110 Marshall-Adams Hall East Lansing, MI 48824-1038 Phone:
517-353-5972 Fax: 517-432-1068
How to check panel unit root with the help of summary statistics
because panel data consist of many panel units (N) and times (T), we can use xtsum command command to summarize variables according to panel dataset - and we can see that between variations
mean variations across panel units, and within variations mean variations across times within a panel unit,
therefore if within variations are greater than between, it is highly likely that the variables are nonstationary or have unit root
Yes normally. But to eliminate serial correlation from model we may change number of lags...So I can use
lag 1 for Johansen cointegration test but change lag for VECM? Sayed Hossain Guideline is normally we
need to follow lag selection criteria suggestion to choose lags but if serial correlation still exists then we
can change lag number.
hlo dear friends i have run PANEL VECM , as panel VECM show long run and short run causality
the question is that , can we still run pairwise causality as well.Thanks in advance
According to Sir Muhammad Anees Yes, GC has no dependence on VECM,, means u can run.
Why we use Hausman test?
Noman Arshed Hausman test is actually used to compare consistent versus efficient estimates. It is not
specially designed for ols it can be used any where when we are comparing consistent vs efficient
method. Examples
1) fe v re
2) 2sls v ols
3) mg v pmg Etc.Hlo dear friends, i want to run granger causality ,as i have core purpose to test
directional relationship ,so should i first run regression then granger causality or direct?
ans by sir Muhammad Anees Run VAR and then GCause
proceduere : go quick--group statistic--granger cusality
What is the main difference between static and dynamic models?
Answer:
Abu Subhi
For static models, the dependent variable responds immediately to changes in the independent
variable.Whereas for dynamic models, the dependent variable does not respond instantaneously to a
change in the independent variable during the period in which the change occurs. In other words, the
independent variables have a lagged (dynamic) effect on the dependent variable.
Distributed lag and autoregressive models are general examples for dynamic models
use one less lag in vecm, suppose if we found optimal lags length 4 then use 3 in vecm...
Sayed Hossain Saeed Aas Khan Meo it is actually suggested by a group of econometricians. They argue,
VECM model automatically converts all variables into first differences but VAR estimation can not do it.
So while use VECM, use 1 less lag suggested by optimal one.Hlo Hi friend i want to know impact of
macroeconomic variables effect on stock returns volatility how to do?
Noman Arshed It can be done using
1) moving standard deviation
2) arch garch modelshould i used first difference data or level form data.
Sayed Hossain commented about VAR and VECM model as such >> If you want to use unrestricted VAR
then use differenced data as VAR can not convert variables automatically to first differenced. But if you
want to run restricted VAR that is VECM, use non stationary data that is level data and EVIEWS or STATA
will convert them to first differenced automatically.
What is the difference between U and ?
The terms RESIDUAL and ERROR, even what they represent the same thing but actualy they are not
exactly the same e.g Y=B0 +B1X +U This is PRF and y=b0 +b1x +e this is SRF SO we can say that u is
the population term which is not observeble and e is sample reg term and is observble I.E e=Y-Y^ ,,,,e is
the estimate of u
how to remove serial correlation from panel data?
if cross sections are more than time then add more variables or try trend variable. if time is more than
cross sections then it means variables might be non stationary, check that and use the panel ARDL.
PANEL ARDL
Professor Abu Subhi commented as below>>
Both Pedroni and Kao cointegration tests assume all variables are integrated of order one. You could try
Dynamic Fixed Effects (DFE) as your short and long run coefficients may be homogeneous across
individuals.
Note:
MG allows short and long run coefficients to differ across individuals. PMG only allows short run
coefficients to differ while the long run are assumed to be homogeneous across individuals.
what is difference between VAR, SVAR, ARCH
Muhammad Anees If I have a model with three variables like x1, x2 and x3 and would like to estimate it
with VAR, then the regression system when each of X will be once a DepVar and others as IndepVars, it
can be considered as simultaneous equations system estimated through vector auto regression
approach. If this regression becomes different between two time spans for the same sample like before
1980s, it behaved differently than after 1980s while estimation approach is still VAR, we can consider it as
Structural VAR. The third approach becomes slighly different. Consider I have to model Y only. If the
residuals of this model becomes of specific structure that its standard deviations and means behave like
autocorrelation is there, the estimation of such models can be through ARCH.
Hlo friends what is difference between chow test and multiple break point test.?
Both are the same, the different between them that Chow test assume that the break dates (points) are
already known.
halo friends, I have check structural brake in my data , I have data from 1980-2014 but confusion is
that test showed three years 2003 ,2007, and 2008 , so how to give value to dummy variables.use
three dummies for the three possible structural break (years), the dummy that is significant in your
regression indicate a structural break at that year. for example if the dummy of the year 2007 is
significant, year 2003 is not significant and the dummy of the year 2008 is significant, then you have
structural break at 2007 and 2008 BUT there is no structural brek at year 2003.
sir i have cheacked multiple breaks one by one in all vriables and i found 3 break in first varibale ,
2 in second ,and 1 in third so how many dummy variables i should use as independent
Then you add 6 dummies in your model and regress, after regressing if there any dummy that is
insignificant drop it and re estimate your model again.
What does it mean when I find that there is cointegration among variables but then I get the long run
coeffients, they turn out to be not significant. Cointegration means that there is a long run relationship but
i find them not significant for my variables
Some time could be the case that there exist a cointegration among variables, but it is just by CHANCE !!
in addition be sure that your model is correct specified.
BUt the series should be stationary and not autocorrelated. But even after taking the first difference you
need to be sure that there is no auto correlation. The auto corr. is important in ARDL not because the
standard errors and so one as we know normally , but in case of ARDL if there is auot corr. the estimated
coefficients will be biased.You add trend to ensure that you not get a spurios regression.
VAR models tells bout long run or short run?
VAR model tells about short run results, How Do I Interpret the P-Values in Linear Regression Analysis?
The p-value for each term tests the null hypothesis that the coefficient is equal to zero (no effect). A low pvalue (< 0.05) indicates that you can reject the null hypothesis. In other words, a predictor that has a low
p-value is likely to be a meaningful addition to your model because changes in the predictor's value are
related to changes in the response variable.
Conversely, a larger (insignificant) p-value suggests that changes in the predictor are not associated with
changes in the response.
In the output below, we can see that the predictor variables of South and North are significant because
both of their p-values are 0.000. However, the p-value for East (0.092) is greater than the common alpha
level of 0.05, which indicates that it is not statistically significant.
In general, multicollinearity is a problem in any econometric framework. For instance, it causes standard
errors to be large even if the underlying specification is correct (in which case the estimates will be
unbiased). The risk of Type II error is thus increased. There are no hard and fast rules on how to deal with
it but possible solutions include excluding, combining, centering or standardizing the problematic
variables or just obtaining more data.That's right. In general, it is practical to accept values below or
around 5 as fine when the sample size is relatively small, and values below or around 10 as fine for
relatively large samples. What is a large or a small sample is totally up to you. In your specific case, 5.16
might be just acceptable.
Dear Josef
You do not need to report the results of all the cointegrating equations. You should only take the first
equation where only dependent variable is normalized. The number of cointegrating equations only tell
you that there is long run relationship.andThe error correcting coefficients are your adjustment coefficients
.
Kerry Green McCullough You should include all the co-integrating relationships in a multivariate
VECM...that is, if Trace and ME say there are 5 relationships, these should be in the VECM together
these describe what is going on in the model. This is because the the vector of disequilibrium terms
contains one for each co-integrating vector (Alexander, 2008, Vol.2, pg. 246). Asteriou and Hall (2011 pg
378) specifically note that the number of cointegrating equations including in the VECM is determined in
the previous step (Trace and ME tests). In the worked example by Brooks (2008 pg 373) he notes that
keeping the number of cointegrating equations to 1 (when Trace and ME have found more) is done in that
particular example context only for simplicity. You could use Brooks as a reference supporting starting
with 1 for simplicity's sake. If you come across any other references for why one may use 1 rather than
the number found in Trace and ME I would be very appreciative if you were to share the references....I
would love to learn more about this approach and Brooks does not give a full set of reasons for this
method.
T test and F test
It is like t test and F test. t test talk about individual significancy while F test talk about joint significancy.
Suppose X1 is alone significant as per t statistics but X1 and X2 are not jointly significant as per F
statistics.
what if BOUND TEST value F value fall between upper bound and lower bound
You pointed interesting thing. Actually you are right. According to Banrjee et all, pesaran et all,
Muhammad Amine if F-statisic is between the upper bound and lower bound. Then you need to rely on
error correction term. If it is significant and negative, then you have long run connection among the
variables
Difference between Var and granger causality .
Granger causality is JUST (JUST) a F-TEST of the coefficients obtained from VAR model. Telling us if the
the coefficients of X and lags of X are jointly(F-TEST) causing Y AND if the coefficients og Y and all lags
of Y are jointly(F-TEST) causing X. When you use STATA for example, you can not use the command og
granger before using the command of VAR!! and when you use the command of granger you get a table
with p-values; this table as i mentioned is JUST a table of F-TEST of the coefficients in VAR model. Good
Luck
what is meaning of giving a shock
That you can say a dhakka-start to the economic system in question like we normally see the new taxes
in Pakistan which acts like a shock for the whole economy and every sector gets affected.
Why we use adjusted R^2
Normally, R^2 increases when add more variables, but that does not necessary means that R^2 is high
because your model is very good. To solve this problem we consider/use adjusted R^2 that adds panelty
for any additional variable. This gives correct picture/information about how good is your model to explain
the dependent variable.
What is R and R2
Dear Naina Baloch and all others:The difference between R-squared and Adj. R-squared is that the latter
is adjusted for degrees of freedom while R2 is not.In a multiple regression model whenever you add
another explanatory variable in the model, R2 will always increase with the inclusion of every additional
variable in the model but the Adj.R2 may or may not increase. If the additional variable does not add
anything to the explanatory power of the model Adj. R2 will decrease although R2 will still increase. If the
additional variable also adds to the explanatory power of the model then both R2 and Adj-R2 will
increase. In order to compare various models we have to use Adj.R-squared instead of R-squared. Since
different models have different degrees of freedom we can't use simple R2 to compare those model. So
we will have to use Adj.R-squared because it is adjusted for degrees of freedom.
What if I have expect coefficient but residual not normally Distributed and having HSK
Even the residual is not normally distributed we can accept the model as OLS estimators are still BLUE.
Bit if there is heteroscedsaticity we can not accept the model as OLS estimators are no longer bLUE.
Why we take log
If you convert the variable into natural log and run the model, probably heteroscedasticity problem will not
arise in the regression model. it is useful also for the interpretation, if you use log on both sides, then the
coefficients indicates the elasticity. Also, log-transformations of dependent variables greatly reduce the
variances and skewness and kurtosis statistics (see Cameron and Trivedi 2010). Log-scale informs on
relative changes (multiplicative), while linear-scale informs on absolute changes (additive). When you
care about relative changes, use the log-scale; when you care about absolute changes, use linear-scale.
In log transformation you use natural logs of the values of the variable rather than the original raw values.
Log transformation works for data where you can see that the residuals get bigger for bigger values of the
dependent variableECM OR EVEC
Normally I convert all variables into natural log so that heteroscedasticity problem may not arise after
running regression model to remove serial correlation from model.
Engle Granger or Johansen test
When two variables you can use ether Engle Granger or Johansen test but when more than two variables
use Johansen only.
Model must be backed
One other serious problem which I observe while I check M.Phil and Ph.D student synopses, theses and
dissertations is that most of the times they write equations and models to be estimated which are not
backed or supported by economic theory, logic or inuition. As many as 50-70 % of the students (not well
guided by their respective advisers) have equations in their theses/synopses not based on economic
theory. As such, I call such equations hanging in the air with any support. So what you need to do?
Whatever you want to do, you must read relevant literature and try to dig out what economic theory says
about your topic or relationship among your variables of interest. This is very important. So, Economic
theory comes first, and empirical testing and evidence come next. If Economic theory is silent about your
topic of interest, you can find out what other learned researchers and economists have reported about
such research. Literature review helps you in this regard. It may help you word your objectives,
hypotheses and even defining your methodology and econometric model as well. I request all teachers to
correctly guide your students. After all, your students are your ambassadors and they will propagate you.
If you help of them they will portray you in a positive way and vice versa.
For those of you who analyse time series data for your thesis research. Keep in mind that time series data
are usually non-stationary at level. So before modeling your data, you must check the time series for unit
root. If your time series variable are stationary at level, then you may go for OLS estimation. If your
variables are non-stationary at level but stationary at first difference that is it is I(1), then check
cointegration. If there is co-integration among your series based on Johansen Cointegration test, then use
VECM. If there is no conintegration among your variables, the do VAR analysis. The coefficients will then
be of short term rather than long term. If some of your series are I(0) and some are I(1), then you will use
Autoregressive Distributed Lag model
What if data not stationary even 1ST difference
You must apply other test like we normally use Dickey fuller test, but If we have trending data we must
chose other tests
ECM notce
If ECM is +ve means no convergence to equilibrium b/c it shows the the shortrun nature of the variables
as well as how they move towards longrunECM term should be negative to say that model is stable and it
converges to long term. EC term should be negative and statistically sig at the same time
Random of fixed effect model
dear Sara Nasir.... both models are static models....and normally uses when time period less or short
period....by fixed effect model we means residual of model is fixed for all cross section while in randome
effect model residual of model is changes with cross sections.. dear Sara Nasir......there are many things
which can be explained....like fixed effect model have individuals intercept for in regression equation while
in random effect model....different residuals......in fixed effect model it can also be one-way fixed effect
model and two-way fixed effect model...it can be discus..
What if R square is listen sixty percent
If R SQURE is less than sixty but if F statistics is significant then we conclude that our model is accept
able and its thumb of rule.
The properties of ARDL
The bounds test approach to cointegration has certain econometric advantages in comparison to other
single equation cointegration procedures. As pointed out by Emran et al. (2007), the bounds test
approach to cointegration is preferred to other conventional cointegration tests because Monte Carlo
evidence shows that it has several important advantages over other conventional tests. The approach
effectively corrects for a possible endogeneity of explanatory variables and the estimates derived from the
approach exhibit desirable small sample properties. Another important advantage of the ARDL approach
is that one can avoid the uncertainties created by unit root pre-testing as the test can be applied
regardless of whether the series are I(0) or I(1). An added bonus of this approach is that unlike other
conventional tests for cointegration, it can be applied to studies that have a small sample size (Narayan,
2005).
What is the acceptable range of skewness and kurtosis for normal distribution of data?
The values for asymmetry and kurtosis between -2 and +2 are considered acceptable in order to prove
normal univariate distribution (George & Mallery, 2010). George, D., & Mallery, M. (2010). SPSS for
Windows Step by Step: A Simple Guide and Reference, 17.0 update (10a ed.) Boston: Pearson.The
acceptable range for skewness or kurtosis below +1.5 and above -1.5 (Tabachnick & Fidell, 2013). If not,
you have to consider transferring data and considering outliers.
What must be Reliability value of variables Cronbach's Alpha
A scale is reliable if the coefficient value is more than 0.600 (Hair et al., 1998).
What must be the value of normality of the data Skewness and kurtosis
According to George and Mallery (2006, p. 99), a skewness or kurtosis value between +1.0 is regarded
as an excellent value and hence, the data for this study is normally distributed.Kurtosis is a measure of
the peakedness or flatness of a distribution. A kurtosis value near 0 indicates a distribution shape close
to normal. A negative kurtosis indicates a shape flatter than normal, and a positive value indicates more
peaked than normal. An extreme kurtosis (e.g. |k| > 5.0) indicates a distribution where more of the values
are in the tails of the distribution than around the mean. A kurtosis value between +1 is considered
excellent, but a value between +2 is acceptable in many analyses in the life sciences
As kurtosis the value of skewness between +1 is considered excellent, but a value between +2 is
acceptable in many analyses in the life sciences
Degrees of Freedoms
: Let me briefly say that the term, degrees of freedom refers to the number of (logically) independent
pieces of information in a sample of data. Let me offer a simple example, suppose that we have a sample
in which there are five values {3, 6, 8, 10, 13} whose sum is 40. In this particular example I can freely
choose 4 values but the fifth one must be such that it will make the sum of these five numbers equal to
40. Thus there are n-1 = 5-1= 4 degrees of freedom. The value of the unknown fifth sample value is
implicitly being determined from the other four values, and the constraint. That is, once the constraint is
introduced, there are only four logically independent pieces of information in the sample. Thats to say,
there are only four "degrees of freedom", once the sample total is revealed. Spencer Sll and Laxmi
Narayan
Great detail of model selection on the base of data
let's suppose that we have a set of time-series variables, and we want to model the relationship between
them, taking into account any unit roots and/or cointegration associated with the data. First, note that
there are three straightforward situations that we're going to put to one side, because they can be dealt
with in standard ways:1. We know that all of the series are I(0), and hence stationary. In this case, we can
simply model the data in their levels, using OLS estimation, for example.
2. We know that all of the series are integrated of the same order (e.g., I(1)), but they are not
cointegrated. In this case, we can just (appropriately) difference each series, and estimate a standard
regression model using OLS.
3. We know that all of the series are integrated of the same order, and they are cointegrated. In this case,
we can estimate two types of models: (i) An OLS regression model using the levels of the data. This will
provide the long-run equilibrating relationship between the variables. (ii) An error-correction model (ECM),
estimated by OLS. This model will represent the short-run dynamics of the relationship between the
variables.
1. Now, let's return to the more complicated situation mentioned above. Some of the variables in question
may bestationary, some may be I(1) or even fractionally integrated, and there is also the possibility of
cointegration among some of the I(1) variables. In other words, things just aren't as "clear cut" as in the
three situations noted above.
Cautions for ARDL
We need a road map to help us. Here are the basic steps that we're going to follow (with details to be
added below):
Make sure than none of the variables are I(2), as such data will invalidate the methodology.
Formulate an "unrestricted" error-correction model (ECM). This will be a particular type of ARDL model.
Determine the appropriate lag structure for the model in step 2.
Make sure that the errors of this model are serially independent.
Make sure that the model is "dynamically stable".
Perform a "Bounds Test" to see if there is evidence of a long-run relationship between the variables.
If the outcome at step 6 is positive, estimate a long-run "levels model", as well as a separate
"restricted" ECM.
Use the results of the models estimated in step 7 to measure short-run dynamic effects, and the longrun equilibrating relationship between the variables.
Endogenous
Endogenous variable: A factor in a causal model or causal system whose value is determined by the
states of other variables in the system; contrasted with an exogenous variable
Exogenous
Exogenous variable (see also endogenous variable): A factor in a causal model or causal system whose
value is independent from the states of other variables in the system; a factor whose value is determined
by factors or variables outside the causal system under stud
Stochastic
Stochastic variable is a variable whose value is subject to variations due to chance (i.e. randomness, in a
mathematical sense).
Monotonically
Asymptotic
The term asymptotic means approaching a value or curve arbitrarily closely (i.e., as some sort of limit is
taken). A line or curve that is asymptotic to given curve is called the asymptote of
Autoregressive model.
If the model includes one or more lagged values of the dependent variable among
Its explanatory variables, it is called an autoregressive model