Econometrics Lecture Note Chapter 3
Econometrics Lecture Note Chapter 3
PART 2
MULTIPLE REGRESSION ANALYSIS
Assumptions of the Multiple Linear Regression
Each econometric method that would be used for estimation purpose has its own assumptions.
Knowing the assumptions and their consequence if they are not maintained is very important
for the econometrician. In the previous section, there are certain assumptions underlying the
multiple regression model, under the method of ordinary least squares (OLS). Let us see them
one by one.
Assumption 2: Zero mean of u i - the random variable u i has a zero mean for each value of
X i i.e. E(ui ) 0
Assumption 3: Homoscedasticity of the random term - the random term u i has constant
variance. In other words, the variance of each u i is the same for all the X i values.
E(ui2 ) u2 Cons tan t
Assumption 4: Normality of u i - the values of each u i are normally distributed ui N (0, u2 )
Assumption 5: No autocorrelation or serial independence of the random terms - the
successive values of the random term are not strongly correlated. The values of
u i (corresponding to xi ) are independent of the values of any other u j (corresponding to X j ).
E(ui u j ) 0 for i j
Assumption 6: Independence of u i and X i - every disturbance term u i is independent of the
explanatory variables. E (u i X 1i ) E (u i X 2i ) 0
Assumption 7: No errors of measurement in the X ' s - the explanatory variables are measured
without error.
Assumption 8: No perfect multicollinearity among the X ' s - the explanatory variables are not
perfectly linearly correlated.
Assumption 9: Correct specification of the model - the model has no specification error in that
all the important explanatory variables appear explicitly in the function and the mathematical
form is correctly defined (linear or non-linear form and the number of equations in the
model).
i 1
0
i
3.6
ˆ1
^
1
n
(Y ˆ0 ˆ1 X 1 ˆ 2 X 2 ) 2
ei2
i 1
0 3.7
ˆ 2
^
2
Solving equations (3.5), (3.6) and (3.7) simultaneously, we obtain the system of normal
equations given as follows:
^ ^ ^
Y n X X
i 0 1 1i 2 2i 3.8
^ ^ ^
X Y X X X X
1i i 0 1i 1 1
2
2 1i 2i 3.9
^ ^ ^
X Y X X X b X
2i i 0 2i 2 1i 2i 2
2
2i 3.10
Then, letting
yi Yi Y 3.13
The above three equations (3.8), (3.9) and (3.10) can be solved using Matrix operations or
simultaneously to obtain the following estimates:
^ x y x x y x x
2
1
1 1 2 2 1 2
3.14
x x x x
2
1
2
2 1 2
2
^ x y x x y x x
2
2
2 1 1 1 2
3.15
x x x x
2
1
2
2 1 2
2
^
0 Y ˆ1 X 1 ˆ 2 X 2 3.16
Like in the case of simple linear regression, the standard errors of the coefficients are vital in
statistical inferences about the coefficients. We use standard the error of a coefficient to
construct confidence interval estimate for the population regression coefficient and to test the
significance of the variable to which the coefficient is attached in determining the dependent
variable in the model. In this section, we will see these standard errors. The standard error of a
coefficient is the positive square root of the variance of the coefficient. Thus, we start with
defining the variances of the coefficients.
2
2
X 1 x 2 X 2 x12 2 X 1 X 2 x1 x 2
2
^ ^
2
Var 0 ui 1 3.17
n
x1 x2 ( x1 x2 )
2 2 2
^
Variance of 2
^ ^ 2
Var ( 2 ) u
x12 3.19
2
x1 x 2 ( x1 x 2 )
2 2
Where,
^
u2
e 2
i
3.20
n3
Equation 3.20 here gives the estimate of the variance of the random term. Then, the standard
errors are computed as follows:
^
Standard error of 0
^ ^
SE ( 0 ) Var ( 0 ) 3.21
^
Standard error of 1
^ ^
SE( 1 ) Var( 1 ) 3.22
^
Standard error of 2
^ ^
SE( 2 ) Var( 2 ) 3.23
Note: The OLS estimators of the multiple regression model have properties which are parallel
to those of the two-variable model.
R2
y2 3.24
y 2
^ ^
Or R can also be given in terms of the slope coefficients 1 and 2 as :
2
^ ^
1 x1 y 2 x 2 y
R
2
3.25
y 2
In simple linear regression, the higher the R 2 means the better the model is determined by the
explanatory variable in the model. In multiple linear regression, however, every time we
insert additional explanatory variable in the model, the R 2 increases irrespective of the
improvement in the goodness-of- fit of the model. That means high R 2 may not imply that the
model is good.
Thus, we adjust the R 2 as follows:
(n 1)
2
Rady 1 (1 R 2 ) 3.26
(n k )
Where, k = the number of explanatory variables in the model.
In multiple linear regression, therefore, we better interpret the adjusted R 2 than the ordinary
or the unadjusted R 2 . We have known that the value of R 2 is always between zero and one.
But the adjusted R 2 can lie outside this range even to be negative.
In the case of simple linear regression, R 2 is the square of linear correlation coefficient.
Again as the correlation coefficient lies between -1 and +1, the coefficient of determination
( R 2 ) lies between 0 and 1. The R 2 of multiple linear regression also lies between 0 and +1.
The adjusted R 2 , however, can sometimes be negative when the goodness of fit is poor. When
the adjusted R 2 value is negative, we considered it as zero and interpret as no variation of the
dependent variable is explained by regressors.
Please recall that 100(1- ) % confidence interval for i is given as ˆi t / 2,n k se( ˆi ) where
k is the number of parameters to be estimated or the number of variables (both dependent and
explanatory)
Interpretation of the confidence interval: Values of the parameter lying in the interval are
plausible with 100(1- ) % confidence.
H 0 : ˆ1 0 H 0 : ˆ 2 0 H 0 : ˆ K 0
a) b)
H : ˆ 0
1 1 H 1 : ˆ 2 0 H 1 : ˆ K 0
In a) we will like to test the hypothesis that X1 has no linear influence on Y holding other
variables constant. In b) we test the hypothesis that X2 has no linear relationship with Y
holding other factors constant. The above hypotheses will lead us to a two-tailed test however,
one-tailed test might also be important. There are two methods for testing significance of
individual regression coefficients.
a) Standard Error Test: Using the standard error test we can test the above hypothesis.
Thus the decision rule is based on the relationship between the numerical value of the
parameter and the standard error of the same.
1 ˆ
(i) If S ( ˆ i ) i , we accept the null hypothesis, i.e. the estimate of i is not statistically
2
significant.
Conclusion: The coefficient ( ˆ i ) is not statistically significant. In other words, it does not
have a significant influence on the dependent variable.
(ii) If S ( ˆ i ) 1 ˆ i , we fail to accept H0, i.e., we reject the null hypothesis in favour of the
2
alternative hypothesis meaning the estimate of i has a significant influence on the dependent
variable.
Otherwise, reject the null hypothesis. Rejecting H 0 means, the coefficient being tested is
significantly different from 0. Not rejecting H 0 , on the other hand, means we don’ t have
sufficient evidence to conclude that the coefficient is different from 0.
Testing the Overall Significance of Regression Model
Here, we are interested to test the overall significance of the observed or estimated regression
line, that is, whether the dependent variable is linearly related to all of the explanatory
variables. Hypotheses of such type are often called joint hypotheses. Testing the overall
significance of the model means testing the null hypothesis that none of the explanatory
variables in the model significantly determine the changes in the dependent variable. Put in
other words, it means testing the null hypothesis that none of the explanatory variables
significantly explain the dependent variable in the model. This can be stated as:
H 0 : 1 2 0
H 1 : i 0, at least for one i.
The test statistic for this test is given by:
yˆ 2
Fcal k 1
e2
nk
Where, k is the number of explanatory variables in the model.
The results of the overall significance test of a model are summarized in the analysis of
variance (ANOVA) table as follows.
Source of Sum of squares Degrees of Mean sum of Fcal
variation freedom squares
Regression ^2 k 1 ^
MSE
SSE y MSE
y 2
F
k 1 MSR
SSR e 2 nk e 2
Residual
MSR
nk
Total SST y 2
n 1
R2 1
e 2
Hence, e 2
1 R 2 which means e 2
(1 R 2 ) y 2
y 2
y 2
Fcal
y 2
k 1
e2
nk
R2 y2 R2 y2 (n k )
Fcal .
k 1 k 1 (1 R 2 ) y 2
(1 R 2 ) y 2
nk
(n k ) R2
Fcal .
k 1 (1 R 2 )
That means the calculated F can also be expressed in terms of the coefficient of
determination.
Testing the Equality of two Regression Coefficients
Given the multiple regression equation:
Yi 0 1 X 1i 2 X 2i 3 X 3i ... K X Ki U i
We would like to test the hypothesis:
H 0 : 1 2 or 1 2 0 vs. H 1 : H 0 is not true
The null hypothesis says that the two slope coefficients are equal.
Example: If Y is quantity demanded of a commodity, X1 is the price of the commodity and
X2 is income of the consumer. The hypothesis suggests that the price and income elasticity of
demand are the same.
We can test the null hypothesis using the classical assumption that
ˆ 2 ˆ1
t ~ t distribution with N - K degrees of freedom.
SE( ˆ 2 ˆ1 )
Where K = the total number of parameters estimated.
The SE( ˆ 2 ˆ1 ) is given as SE( ˆ 2 ˆ1 ) Var ( ˆ 2 ) Var ( ˆ1 ) 2 cov( 2 , 1 )
Thus the t-statistic is:
ˆ 2 ˆ1
t
Var ( ˆ 2 ) Var ( ˆ1 ) 2 cov( 2 , 1 )
Note: Using similar procedures one can also test linear equality restrictions, for example
1 2 1 and other restrictions.
a) Estimate the coefficients of the economic relationship and fit the model.
To estimate the coefficients of the economic relationship, we compute the entries given in
Table 2
Y 639
Y 53.25
n 12
X 1
3679
X1 306.5833
n 12
X 2
1684
X2 140.3333
n 12
The summary results in deviation forms are then given by:
x 49206.92 x 2500.667
2 2
1 2
x y 1043.25
1 x 2 y 509
x x 1 2 960.6667 y 2
1536.25
1
1 1 2 2 1 2
x x x x
2
1
2
2 1 2
2
(49206.92)(2500.667)- (960.667) 123050121 922880.51 2
3097800.2
0.025365
122127241
2
2 1 1 1 2
x x x x
2
1
2
2 1 2
2
(49206.92)(2500.667)- (960.667) 123050121 922880.51 2
- 26048538
0.21329
122127241
^
2
e 2
i
1401.223 1401.223
155.69143
n3 12 3
u
9
^
Variance of 1
x1 x2 ( x1 x2 )
2 2
12212724
^
Standard error of 1
^ ^
SE( 1 ) Var ( 1 ) 0.003188 0.056462
^
Variance of 2
^
Var ( 2 ) u
^ 2 x12
155.69143(
49206.92
) 0.0627
2
x1 x2 ( x1 x 2 )
2 2
122127241
^
Standard error of 2
^ ^
SE( 2 ) Var ( 2 ) 0.0627 0.25046
Similarly, the standard error of the intercept is found to be 37.98177. The detail is left for you
as an exercise.
c) Calculate and interpret the coefficient of determination.
We can use the following summary results to obtain the R2.
yˆ 2
135.0262
e 2
1401.223
y 2
1536.25 (The sum of the above two). Then,
^ ^
1 x1 y 2 x 2 y (0.025365)(1043.25) (-0.21329)(-509)
R
2
0.087894
y 2
1356.25
e 2
1401.223
or R 2 1 1 0.087894
1356.25
y 2
The critical value (t 0.05, 9) to be used here is 2.262. Like the standard error test, the t- test
revealed that both X1 and X2 are insignificant to determine the change in Y since the
calculated t values are both less than the critical value.
Exercise: Test the significance of X1 and X2 in determining the changes in Y using the
standard error test.
g) Test the overall significance of the model. (Hint: use = 0.05)
This involves testing whether at least one of the two variables X1 and X2 determine the
changes in Y. The hypothesis to be tested is given by:
H 0 : 1 2 0
H 1 : i 0, at least for one i.
The ANOVA table for the test is give as follows:
Source of Sum of Squares Degrees of Mean Sum of Squares Fcal
variation freedom
Regression ^ 2 ^ MSR
k 1 =3-1=2 MSR 135.0262
SSR y 135.0262 y 2
F
MSE
k 1 2 0.433634
67.51309
Residual SSE e 1401.223
2
n k =12- e 2 1401.223
MSE
3=9 nk 9
155.614
Total SST y 1536.25 n 1
2 =12-
1=11
In this case, the calculated F value (0.4336) is less than the tabulated value (3.98). Hence, we
do not reject the null hypothesis and conclude that there is no significant contribution of the
variables X1 and X2 to the changes in Y.
h) Compute the F value using the R2.
From the above fitted model, we can see that mean salary of public school teachers in the
West is about $26,158.62. The mean salary of teachers in the Northeast is lower by $1734.47
than those of the West and those in the South is lower by $3264.42. Doing this, we will find
the average salaries in the latter two regions are about $24,424 and $22,894.
In order to know the statistical significance of the mean salary differences, we can run the
tests we have discussed in previous sections. The other results can also be interpreted the way
we discussed previously.
Y
X => Y
K
X
Output
Input
Thus, given the assumption of a constant elasticity, the proper form is the exponential (log-
linear) form.
Given: Yi 0 X i i eU i
The log-linear functional form for the above equation can be obtained by a logarithmic
transformation of the equation.
ln Yi ln 0 i ln X i U i
The model can be estimated by OLS if the basic assumptions are fulfilled.
ln Yi ln 0 1 ln X i
1
Yi 0 X i
The model is also called a constant elasticity model because the coefficient of elasticity
between Y and X (1) remains constant.
Y X d ln Y
1
X Y d ln X
This functional form is used in the estimation of demand and production functions.
Note: We should make sure that there are no negative or zero observations in the data set
before we decide to use the log-linear model. Thus log-linear models should be run only if all
the variables take on positive values.
c) Semi-log Form
The semi-log functional form is a variant of the log-linear equation in which some but not all
of the variables (dependent and independent) are expressed in terms of their logs. Such
models expressed as:
( i ) Yi 0 1 ln X 1i U i ( lin-log model ) and ( ii ) ln Yi 0 1 X 1i U i ( log-lin
model ) are called semi-log models. The semi-log functional form, in the case of taking the
log of one of the independent variables, can be used to depict a situation in which the impact
of X on Y is expected to ‘ tail off’ as X gets bigger as long as 1 is greater than zero.
1<0
Y=0+1Xi
1>0
Example: The Engel’ s curve tends to flatten out, because as incomes get higher, a smaller
percentage of income goes to consumption and a greater percentage goes to saving.
Consumption thus increases at a decreasing rate.
Growth models are examples of semi-log forms
d) Polynomial Form
Polynomial functional forms express Y as a function of independent variables some of which
are raised to powers others than one. For example in a second degree polynomial (quadratic)
equation, at least one independent variable is squared.
Y 0 1 X 1i 2 X 1i 3 X 2i U i
2
Such models produce slopes that change as the independent variables change. Thus the slopes
of Y with respect to the Xs are
Y Y
1 2 2 X 1 , and 3
X 1 X 2
In most cost functions, the slope of the cost curve changes as output changes.
A) B)
X
Xi Impact of age on earnings
a typical cost curve
Simple transformation of the polynomial could enable us to use the OLS method to estimate
the parameters of the model
X1 X 3
2
Setting
Y 0 1 X 1i 2 X 3 3 X 2i U i
e) Reciprocal Transformation (Inverse Functional Forms)
The inverse functional form expresses Y as a function of the reciprocal (or inverse) of one or
more of the independent variables (in this case X1):
1
Yi 0 1 ( ) 2 X 2i U i
X 1i
Or
1
Yi 0 1 ( ) 2 X 2i U i
X 1i
The reciprocal form should be used when the impact of a particular independent variable is
expected to approach zero as that independent variable increases and eventually approaches
infinity. Thus as X1 gets larger, its impact on Y decreases.
1 0 0
Y 0
X 1i 1 0
0
1 0 0
Y 0
X 1i 1 0
ECONOMETRIC PROBLEMS
Pre-test Questions
1. What are the major CLRM assumptions?
2. What happens to the properties of the OLS estimators if one or more of these
assumptions are violated, i.e. not fulfilled?
3. How we can check if an assumption is violated or not?
Assumptions Revisited
In many practical cases, two major problems arise in applying the classical linear regression
model.
1) those due to assumptions about the specification of the model and about the
disturbances and
2) those due to assumptions about the data
The following assumptions fall in either of the categories.
The regression model is linear in parameters.
The values of the explanatory variables are fixed in repeated sampling (non-
stochastic).
The mean of the disturbance (ui) is zero for any given value of X i.e. E(ui) = 0
The variance of ui is constant i.e. homoscedastic
There is no autocorrelation in the disturbance terms
The explanatory variables are independently distributed with the ui.
The number of observations must be greater than the number of explanatory
variables.
There must be sufficient variability in the values taken by the explanatory
variables.
There is no linear relationship (multicollinearity) among the explanatory variables.
The stochastic (disturbance) term ui are normally distributed i.e., ui ~ N(0, ²)
The regression model is correctly specified i.e., no specification error.
With these assumptions we can show that OLS are BLUE, and normally distributed. Hence it
was possible to test Hypothesis about the parameters. However, if any of such assumption is
relaxed, the OLS might not work. We shall not examine in detail the violation of some of the
assumptions.
Violations of Assumptions
The Zero Mean Assumption i.e. E(ui)=0
Consequences of Heteroscedasticity
If the error terms of an equation are heteroscedastic, there are three major consequences.
a) The ordinary least square estimators are still linear since heteroscedasticity does not
cause bias in the coefficient estimates. The least square estimators are still unbiased.
b) Heteroscedasticity increases the variance of the partial regression coefficients but it
does affect the minimum variance property. Thus, the OLS estimators are inefficient.
Thus the test statistics – t-test and F-test – cannot be relied on in the face of
uncorrected heteroscedasticity.
Detection of Heteroscedasticity
There are no hard and fast rules (universally agreed upon methods) for detecting the presence
of heteroscedasticity. But some rules of thumb can be suggested. Most of these methods are
based on the examination of the OLS residuals, ei, since these are the once we observe and
not the disturbance term ui. There are informal and formal methods of detecting
heteroscedasticity.
a) Nature of the problem
In cross-sectional studies involving heterogeneous units, heteroscedasticity is the rule rather
than the exception.
Example: In small, medium and large sized agribusiness firms in a study of input expenditure
in relation to sales, the rate of interest, etc. heteroscedasticity is expected.
b) Graphical method
If there is no a priori or empirical information about the nature of heteroscedasticity, one
could do an examination of the estimated residual squared, ei² to see if they exhibit any
systematic pattern. The squared residuals can be plotted either against Y or against one of the
explanatory variables. If there appears any systematic pattern, heteroscedasticity might exist.
These two methods are informal methods.
c) Park Test
Park suggested a statistical test for heteroscedasticity based on the assumption that the
variance of the disturbance term (i²) is some function of the explanatory variable Xi.
Park suggested a functional form as: i 2 X i e vi which can be transferred to a linear
2
A high rank correlation suggests the presence of heteroscedasticity. If more than one
explanatory variable, compute the rank correlation coefficient between ei and each
explanatory variable separately.
Remedial Measures
OLS estimators are still unbiased even in the presence of heteroscedasticity. But they are not
efficient, not even asymptotically. This lack of efficiency makes the usual hypothesis testing
procedure a dubious exercise. Remedial measures are, therefore, necessary. Generally the
solution is based on some form of transformation.
a) The Weighted Least Squares (WLS)
Given a regression equation model of the form
Yi 0 1 X i U i
The weighted least square method requires running the OLS regression to a transformed data.
The transformation is based on the assumption of the form of heteroscedasticity.
Assumption One: Given the model Yi 0 1 X 1i U i
If var(U i ) i X i , then E (U i ) X 1i
2 2 2 2 2
1
0 ( ) 1 Vi
X 1i
Now E (Vi 2 ) E ( U i ) 1
2
E (U i ) 2
2
X 1i X 1i
Y 1
Hence the variance of Ui is now homoscedastic and regress on .
X 1i X 1i
Assumption Two: Again given the model
E (U i ) 2 2 E (Yi )
2
1 X 1i
0 1 Vi
E (Yi ) E (Yi )
Ui
Again it can be verified that Vi gives us a constant variance ²
E (Yi )
2
Ui
2 E (Yi ) 2
1 1
E (Vi ) E E (U i ) 2
2
E (Yi ) E (Yi ) 2
E (Yi )2
E (U iU j ) 0 , i j
But if there is any interdependence between the disturbance terms then we have
autocorrelation
E(U iU j ) 0 , i j
Causes of Autocorrelation
Serial correlation may occur because of a number of reasons.
Inertia (built in momentum) – a salient feature of most economic variables time series
(such as GDP, GNP, price indices, production, employment etc) is inertia or
sluggishness. Such variables exhibit (business) cycles.
Specification bias – exclusion of important variables or incorrect functional forms
Lags – in a time series regression, value of a variable for a certain period depends on
the variable’ s previous period value.
Manipulation of data – if the raw data is manipulated (extrapolated or interpolated),
autocorrelation might result.
Autocorrelation can be negative as well as positive. The most common kind of serial
correlation is the first order serial correlation. This is the case in which this period error
terms are functions of the previous time period error term.
et-1 et-1
et = -ve
et-1 = -ve
et-1 = +ve
et = -ve
There are more accurate tests for the incidence of autocorrelation. The most common test of
autocorrelation is the Durbin-Watson Test.
The Durbin-Watson d Test
The test for serial correlation that is most widely used is the Durbin-Watson d test. This test is
appropriate only for the first order autoregressive scheme.
U t PU t 1 Et then Et PE t 1 U t
The test may be outlined as
HO : P 0
H1 : P 0
This test is, however, applicable where the underlying assumptions are met:
The regression model includes an intercept term
The serial correlation is first order in nature
The regression does not include the lagged dependent variable as an explanatory
variable
There are no missing observations in the data
The equation for the Durban-Watson d statistic is
N
(e t et 1 ) 2
d t 2
N
e
2
t
t 1
Which is simply the ratio of the sum of squared differences in successive residuals to the RSS
Note that the numerator has one fewer observation than the denominator, because an
observation must be used to calculate et 1 . A great advantage of the d-statistic is that it is
based on the estimated residuals. Thus it is often reported together with R², t, etc.
The d-statistic equals zero if there is extreme positive serial correlation, two if there is no
serial correlation, and four if there is extreme negative correlation.
thus d
(2e ) t
2
and d 4
e
2
t
3. No serial correlation: d 2
d
(e e t t 1 )2
e t
2
et 1 2 et et 1
2
2
e e
2 2
t t
But Durbin and Watson have successfully derived the upper and lower bound so that if the
computed value d lies outside these critical values, a decision can be made regarding the
presence of a positive or negative serial autocorrelation.
Thus
(e e e et 1 2 et et 1
2 2
t 1 )2
d
t t
e e
2 2
t t
2(1
e e t t 1
)
e
2
t 1
ˆ ) since
d 2(1 P
e e t t 1 ˆ
P
e
2
t 1
Reject H0 Reject H0
+ve -ve
autocorr. autocorr.
accept H0
no serial
correlation
0 d
dL dU 4-dU 4-dL 4
Note: Other tests for autocorrelation include the Runs test and the Breusch-Godfrey (BG) test.
There are so many tests of autocorrelation since there is no particular test that has been judged
to be unequivocally best or more powerful in the statistical sense.
Remedial Measures for Autocorrelation
Since in the presence of serial correlation the OLS estimators are inefficient, it is essential to
seek remedial measures.
1) The solution depends on the source of the problem.
The methods discussed above to solve the problem of serial autocorrelation are basically two
step methods. In step 1, we obtain an estimate of the unknown and in step 2, we use that
estimate to transform the variables to estimate the generalized difference equation.
Note: The Cochrane-Orcutt Iterative Method is also another method.
Multicollinearity: Exact linear correlation between Regressors
One of the classical assumptions of the regression model is that the explanatory variables are
uncorrelated. If the assumption that no independent variable is a perfect linear function of one
or more other independent variables is violated we have the problem of multicollinearity. If
the explanatory variables are perfectly linearly correlated, the parameters become
indeterminate. It is impossible to find the numerical values for each parameter and the method
of estimation breaks.
If the correlation coefficient is 0, the variables are called orthogonal; there is no problem of
multicollinearity. Neither of the above two extreme cases is often met. But some degree of
inter-correlation is expected among the explanatory variables, due to the interdependence of
economic variables.
Multicollinearity is not a condition that either exists or does not exist in economic functions,
but rather a phenomenon inherent in most relationships due to the nature of economic
magnitude. But there is no conclusive evidence which suggests that a certain degree of
multicollinearity will affect seriously the parameter estimates.
Reasons for Existence of Multicollinearity
without severe
multicollinearity
with severe
multicollinearity
̂
(3) The computed t-ratios will fall i.e. insignificant t-ratios will be observed in the
presence of multicollinearity. t since SE( ˆ ) increases t-falls. Thus
ˆ
SE ( )
one may increasingly accept the null hypothesis that the relevant true population’ s
value is zero
So the combination of high R² with low calculated t-values for the individual regression
coefficients is an indicator of the possible presence of severe multicollinearity.
Drawback: a non-multicollinear explanatory variable may still have a significant coefficient
even if there is multicollinearity between two or more other explanatory variables Thus,
equations with high levels of multicollinearity will often have one or two regression
coefficients significantly different from zero, thus making the “ high R² low t” rule a poor
indicator in such cases.
1) High pair-wise (simple) correlation coefficients among the regressors (explanatory
variables).
If the R’ s are high in absolute value, then it is highly probable that the X’ s are highly
correlated and that multicollinearity is a potential problem. The question is how high r should
be to suggest multicollinearity. Some suggest that if r is in excess of 0.80, then
multicollinearity could be suspected.