ü The Nature of Heteroscedasticity
There are several reasons why the variances of ui may be a variable, some of which are as follows:
1. Following the error-learning models, as people learn, their errors of behavior become smaller over time.
2. As incomes grow, people have more discretionary income and hence more scope of choice about the disposition of their income.
3. As data collecting techniques, is likely to decrease.
4. Heteroscedasticity can also arise as a result of the presence of outliers.
5. Another source of heteroscedasticity arises from violating Assumption 9 of CLRM, namely, that the regression model is correctly specified.
6. Another source of heteroscedasticity is skewness in the distribution of one or more regressors included in the model.
7. Heteroscedasticity can also arise because of (1) incorrect data transformation and (2) incorrect functional form.
ü OLS Estimation in the Presence of Heteroscedasticity
var(β2) = ∑Xi2
(∑Xi2)2
var(β3) =
∑Xi2
ü The Method of Generalized Least Squares
- Takes such information into account explicitly and is therefore capable of producing estimators that are BLUE.
ü Difference Between OLS and GLS
OLS:
∑ui2 = ∑(Yi – β1 – β2Xi)2
GLS:
∑wiui2 = ∑wi(Yi – β1Xi – β2Xi)2
ü Consequence of Using OLS in the Presence of Heteroscedasticity
ü OLS Estimation Disregarding Heteroscedasticity
In short, if we persist in using the usual testing procedures despite heteroscedasticity, whatever conclusions we draw or inferences we make may be very misleading.
ü Detection of Heteroscedasticity
ü Informal Methods
Nature of the Problem. Very often the nature of the problem under consideration suggests whether heteroscedasticity is likely to be encountered.
Graphical Method. If there are no priori or empirical information about the nature of heteroscedasticity, in practice one can do the regression analysis on the assumption that there is no heteroscedasticity and then do the postmortem examination of the residual squared ui2 to see if they exhibit any systematic pattern.
ü Formal Methods
Park Test. Park formalizes the graphical method by suggesting that is some function of the explanatory variable Xi.
Glejser Test. After obtaining the residuals ui from the OLS regression, Glejser suggests regressing the absolute values of ui on the X variable that is thought to be closely associated with .
Spearman’s Rank Correlation Test.
rs = 1 - ∑di2
n(n2 – 1)
Step 1: Fit the regression to the data on Y and X and obtain the residuals ui.
Step 2: Ignoring the sign of ui, that is, taking their absolute value ui , rank both ui and Xi or (Yi) according to an ascending or descending order and compute the Spearman’s rank correlation coefficient given previously.
Step 3: Assuming that the population rank correlation coefficient ρs is zero and n>8, the significance of the sample rs can be tested by the t test as follows.
t = rs n – 2
1 – r2s
Goldfeld-Quandt Test. This popular method is applicable if one assumes that the heteroscedastic variance , is positively related to one of the explanatory variable in the regression model.
Step 1: Order or rank the observations according to the values of Xi, beginning with the lowest X value.
Step 2: Omit c central observations, where c is specified a priori, and divide the remaining (n – c) observations into two groups each of (n – c) /2 observations.
Step 3: Fit separate OLS regressions to the first (n – c)/2 observations and the last (n – c)/2 observations and obtain the respective residual sums of squares RSS1 and RSS2, RSS1 representing the RSS from the regression corresponding to the smaller Xi values and RSS2 that from the larger Xi values.
Step 4: Compute the ratio
λ = RSS2/df
RSS1/df
If ui are assumed to be normally distributed and if the assumption of homoscedasticity is valid.
Breusch-Pagan-Godfrey Test
Step 1: Estimate
Yi = β1 + β2X2i + βkXk + ui
by OLS and obtain the residuals u1, u2, . . . ,un
Step 2: Obtain = ∑ui2/n
Step 3: Construct variables pi defined as
pi = ui2/
which is simply each residual squared divided by .
Step 4: Regress pi thus considered on the z’s as
pi = α1 + α2Z2i + . . . + αmZmi + vi
where vi is the residual term of this regression.
Step 5: Obtain the ESS and define
= ½(ESS)
White’s General Heteroscedasticity Test
Step 1: Given the data, we estimate
Yi = β1 + β2X2i + β3X3i + ui
and obtain the residuals, ui.
Step 2: We then run the following regression:
ui2 = α1 + α2X2i + α3X3i + α4X2i + α5X3i + α6X2iX3i + vi
Step 3: Under the null hypothesis that there is no heteroscedasticity, it can be shown that the sample size (n) times the R2 obtained from the auxiliary regression asymptotically follows the chi-square distribution with df equal to the number of regressors in the auxiliary regression. That is,
n R2 asy X2df
Step 4: If the chi-square value obtained in n R2 asy X2df exceeds the critical chi-square value at the chosen level of significance, conclusion is that there is heteroscedasticity.
ü Other Tests of Heteroscedasticity
· Koenker-Bassett (KB) Test
ü Remedial Measures
ü When is Known: The Method of Weighted Least Squares
ü When not Known
· Plausible assumptions about heteroscedasticity pattern
Assumption 1: The error variance is proportional to Xi2
E(ui2) = Xi2
Assumption 2: The error variance is proportional to Xi. The square root transformation:
E(ui2) = Xi
Assumption 3: The error variance is proportional to the square of the mean value of Y.
E(ui2) = [E(Yi)]2
Assumption 4: A log transformation such as
lnYi = β1 + β2lnXi + ui
very often reduces heteroscedasticity when compared with the regression Yi = β1 + β2Xi + ui
No comments:
Post a Comment