Skip to main content
Log in

Prediction-based regularization using data augmented regression

  • Published:
Statistics and Computing Aims and scope Submit manuscript

Abstract

The role of regularization is to control fitted model complexity and variance by penalizing (or constraining) models to be in an area of model space that is deemed reasonable, thus facilitating good predictive performance. This is typically achieved by penalizing a parametric or non-parametric representation of the model. In this paper we advocate instead the use of prior knowledge or expectations about the predictions of models for regularization. This has the twofold advantage of allowing a more intuitive interpretation of penalties and priors and explicitly controlling model extrapolation into relevant regions of the feature space. This second point is especially critical in high-dimensional modeling situations, where the curse of dimensionality implies that new prediction points usually require extrapolation. We demonstrate that prediction-based regularization can, in many cases, be stochastically implemented by simply augmenting the dataset with Monte Carlo pseudo-data. We investigate the range of applicability of this implementation. An asymptotic analysis of the performance of Data Augmented Regression (DAR) in parametric and non-parametric linear regression, and in nearest neighbor regression, clarifies the regularizing behavior of DAR. We apply DAR to simulated and real data, and show that it is able to control the variance of extrapolation, while maintaining, and often improving, predictive accuracy.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Abu-Mostafa, Y.: Hints. Neural Comput. 7, 639–671 (1995)

    Article  Google Scholar 

  • Bedrick, E.J., Christensen, R., Johnson, W.: A new perspective on priors for generalized linear models. J. Am. Stat. Assoc. 91, 1450–1460 (1996)

    Article  MathSciNet  MATH  Google Scholar 

  • Bickel, P., Bo, L.: Regularization in statistics. Test, pp. 271–344

  • Bickel, P., Ritov, Y., Tsybakov, A.B.: Simultaneous analysis of Lasso and Dantzig selector. Ann. Stat. 37 (2009)

  • Breiman, L.: Bagging predictors. Mach. Learn. 24(2) (1996)

  • Breiman, L.: Statistical modeling: the two cultures. Stat. Sci. (2001)

  • Breiman, L., Friedman, J., Olshen, R., Stone, C.: Classification and Regression Trees. Wadsworth, Belmont (1984)

    MATH  Google Scholar 

  • Christensen, R.: Analysis of Variance, Design and Regression: Applied Statistical Methods. Chapman and Hall, New York (1996)

    Google Scholar 

  • Dasarathy, B.: Nearest Neighbor Pattern Classification Techniques. IEEE Comput. Soc., Los Alamitos (1991)

    Google Scholar 

  • Harrison, D., Rubinfeld, D.L.: Hedonic prices and the demand for clean air. J. Environ. Econ. Manage. 5, 81–102 (1978)

    Article  MATH  Google Scholar 

  • Hastie, T., Tibshirani, R., Friedman, J.H.: The Elements of Statistical Learning. Springer, New York (2001)

    MATH  Google Scholar 

  • Hoerl, A., Kennard, R.: Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 12(3), 55–67 (1970)

    Article  MATH  Google Scholar 

  • Hooker, G.: Diagnosing extrapolation: Tree-based density estimation. In: Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (2004)

    Google Scholar 

  • Lehmann, EL, Casella, G.: Theory of Point Estimation. Springer, New York (1998)

    MATH  Google Scholar 

  • Mammen, E., van de Geer, S.: Locally adaptive regression splines. Ann. Statist. 25(1), 387–413 (1997)

    Article  MathSciNet  MATH  Google Scholar 

  • Munson, MA, Webb, K., Sheldon, D., Fink, D., Hochachka, W.M., Iliff, M., Riedewald, M., Sorokina, D., Sullivan, B., Wood, C., Kelling, S.: The ebird reference dataset. http://www.avianknowledge.net (2009)

  • Niyogi, P., Girosi, F., Poggio, T.: Incorporating prior information in machine learning by creating virtual examples. Proc. IEEE 86(11), 2196–2209 (1998)

    Article  Google Scholar 

  • Pace, R.K., Barry, R.: Sparse spatial autoregressions. Stat. Probab. Lett. 33, 291–297 (1997)

    Article  MATH  Google Scholar 

  • R Development Core Team: R: a language and environment for statistical computing. R foundation for statistical computing, Vienna, Austria, URL http://www.R-project.org (2007). ISBN 3-900051-07-0

  • Rifkin, R.M., Lippert, R.A.: Value regularization and the fenchel duality. J. Mach. Learn. Res. 8, 441–479 (2007)

    MathSciNet  MATH  Google Scholar 

  • Tibshirani, R.: Regression shrinkage and selection via the lasso. J. R. Stat. Soc. B 58(1), 267–288 (1996)

    MathSciNet  MATH  Google Scholar 

  • Tsutakawa, R.K., Lin, Y.H.: Bayesian estimation of item response curves. Psychometrika 51, 251–267 (1986)

    Article  MathSciNet  MATH  Google Scholar 

  • Vapnik, V.N.: The Nature of Statistical Learning Theory. Springer, Berlin (1996)

    Google Scholar 

  • Vapnik, V.N.: Statistical Learning Theory. Wiley, New York (1998)

    MATH  Google Scholar 

  • Wahba, G.: Spline Models for Observational Data. CBMS-NSF Regional Conference Series in Applied Mathematics (1990)

  • Zhu, J., Hastie, T.: Kernel logistic regression and the import vector machine. J. Comput. Graph. Statist. 14, 185–205 (2005)

    Article  MathSciNet  Google Scholar 

  • Zou, H., Hastie, T.: Regularization and variable selection via the elastic net. J. R. Stat Soc. 67, 301–320 (2005)

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Giles Hooker.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Hooker, G., Rosset, S. Prediction-based regularization using data augmented regression. Stat Comput 22, 237–249 (2012). https://doi.org/10.1007/s11222-010-9220-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11222-010-9220-1

Keywords

Navigation