Abstract
The paper examines robustness of results from cross-sectional regression paying attention to the impact of multicollinearity. It is well known that the reliability of estimators (least-squares or maximum-likelihood) gets worse as the linear relationships between the regressors become more acute. We resolve the discussion in a spatial context, looking closely into the behaviour shown, under several unfavourable conditions, by the most outstanding misspecification tests when collinear variables are added to the regression. A Monte Carlo simulation is performed. The conclusions point to the fact that these statistics react in different ways to the problems posed.
Similar content being viewed by others
References
Anselin L, Florax RJGM (1995) Small sample properties of tests for spatial dependence in regression models. In: Anselin L, Florax R (eds) New directions in spatial econometrics. Springer, Berlin Heidelberg New York, pp 21–74
Belsley D, Kuh E, Welsh R (1980) Regression diagnostics: identifying influential data and sources of collinearity. Wiley, New York
Chatterjee S, Hadi A (1988) Sensitivity analysis in linear regression. Wiley, New York
Draper N, Nostrand R (1979) Ridge regression and James–Stein estimation: review and comments. Technometrics 21:451–465
Farrar D, Glauber R (1967) Multicollinearity in regression analysis: the problem revisited. Rev Econ Stat 49:92–107
Florax RJGM, de Graaff T (2004) The performance of diagnostics tests for spatial dependence in linear regression models: a meta-analysis of simulation studies. In: Anselin L, Florax R, Rey S (eds) Advances in spatial econometrics: methodology, tools and applications. Springer, Berlin Heidelberg New York, pp 29–66
Greene W (2003) Econometric analysis, 5th edn. Prentice Hall, New Jersey
Hocking R (1983) Developments in linear regression methodology: 1959–1982. Technometrics 25:219–230
King B (1969) Comments on “Factor analysis and regression”. Econometrica 37:538–540
Kosfeld R, Lauridsen J (2006) Factor analysis regression. Working Paper, Department of Economics, University of Kassel
Kumar T (1975) Multicollinearity in regression analysis. Rev Econ Stat 57:365–366
Lauridsen J, Mur J (2006) Multicollinearity and outliers in cross-sectional analysis. Working paper, Department of Business and Economics, University of Southern Denmark
Mur J, Lauridsen J (2006) Outliers and spatial dependence in cross-sectional analysis. Environment and Planning A (forthcoming)
O’Hagan J, McCabe B (1975) Tests for the severity of multicollinearity in regression analysis: a comment. Rev Econ Stat 57:368–370
Scott J (1966) Factor analysis and regression. Econometrica 34:552–562
Scott J (1969) Factor analysis and regression revisited. Econometrica 37:719
Wichers C (1975) The detection of multicollinearity: a comment. Rev Econ Stat 57:366–368
Acknowledgments
This work has been carried out with the financial support of project SEC 2002-02350 of the Spanish Ministerio de Educatión. The authors also wish to thank Ana Angulo for her invaluable and disinterested collaboration.
Author information
Authors and Affiliations
Corresponding author
Appendix: Misspecification test used
Appendix: Misspecification test used
The tests used always refer to the model of the null hypothesis; that is, of the static type such as: y = Xβ + u. This model has been estimated by LS, where \(\hat{\sigma}^{2}\) and \(\hat{\beta}\) are the corresponding LS estimations and \(\hat{u}\) the residual series. The tests are the following (see Anselin and Florax 1995; or Florax and de Graaff 2004, for details):
Moran’s I test:
LM-ERR test:
LM-EL test:
KR test:
LM-LAG test:
LM-LE test:
SARMA test:
Moreover, \(R_{j} = T_{1} + \frac{{\hat{\beta}'X'W`MWX\hat{\beta}}}{{\hat{\sigma}^{2}}}\) and M = I−X(X′X)− 1 X′. Furthermore, \(\hat{e}\) is the vector of residuals from the auxiliary regression of the Kelejian–Robinson (KR) test, of order h R × 1, Z is the matrix of exogenous variables included in the last regression and \(\hat{\gamma}\) the estimated coefficients obtained for the corresponding vector of parameters.
As is well-known, the asymptotic distribution of the standardised Moran’s I, obtained as \(\frac{{I - E(I)}}{{{\sqrt {V(I)}}}}\), with \(E(I) = \frac{R}{{S_{0} (R - k)}}{\rm tr}(MW)\) and
is an N(0,1); the two Lagrange Multipliers that follow, LM-ERR and LM-EL, have an asymptotic χ 2 (1), the distribution of the KR test is a χ 2(m), with m being the number of regressors included in the auxiliary regression. The three final tests also have a chi-square distribution, with one degree of freedom in the first two, and two degrees of freedom in the SARMA test.
Rights and permissions
About this article
Cite this article
Lauridsen, J., Mur, J. Multicollinearity in cross-sectional regressions. J Geograph Syst 8, 317–333 (2006). https://doi.org/10.1007/s10109-006-0031-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10109-006-0031-z