Search results
Results from the WOW.Com Content Network
Heteroscedasticity often occurs when there is a large difference among the sizes of the observations. A classic example of heteroscedasticity is that of income versus expenditure on meals. A wealthy person may eat inexpensive food sometimes and expensive food at other times. A poor person will almost always eat inexpensive food.
There should then be a later stage of analysis to examine whether the errors in the predictions from the regression behave in the same way across the dataset. Thus the question becomes one of the homogeneity of the distribution of the residuals, as the explanatory variables change. See regression analysis.
For any non-linear model (for instance logit and probit models), however, heteroskedasticity has more severe consequences: the maximum likelihood estimates of the parameters will be biased (in an unknown direction), as well as inconsistent (unless the likelihood function is modified to correctly take into account the precise form of ...
An alternative to the White test is the Breusch–Pagan test, where the Breusch-Pagan test is designed to detect only linear forms of heteroskedasticity. Under certain conditions and a modification of one of the tests, they can be found to be algebraically equivalent.
Under the classical assumptions, ordinary least squares is the best linear unbiased estimator (BLUE), i.e., it is unbiased and efficient. It remains unbiased under heteroskedasticity, but efficiency is lost. Before deciding upon an estimation method, one may conduct the Breusch–Pagan test to examine the presence of heteroskedasticity.
Weighted least squares (WLS), also known as weighted linear regression, [1] [2] is a generalization of ordinary least squares and linear regression in which knowledge of the unequal variance of observations (heteroscedasticity) is incorporated into the regression.
Notice the relation between the variance and the mean, which implies, for example, heteroscedasticity in a linear model. Therefore, the goal is to find a function g {\displaystyle g} such that Y = g ( X ) {\displaystyle Y=g(X)} has a variance independent (at least approximately) of its expectation.
Glejser test for heteroscedasticity, developed in 1969 by Herbert Glejser, is a statistical test, which regresses the residuals on the explanatory variable that is thought to be related to the heteroscedastic variance. [1]