Skip to main content
Log in

A comparison of methods for the fitting of generalized additive models

  • Published:
Statistics and Computing Aims and scope Submit manuscript

Abstract

There are several procedures for fitting generalized additive models, i.e. regression models for an exponential family response where the influence of each single covariates is assumed to have unknown, potentially non-linear shape. Simulated data are used to compare a smoothing parameter optimization approach for selection of smoothness and of covariates, a stepwise approach, a mixed model approach, and a procedure based on boosting techniques. In particular it is investigated how the performance of procedures is linked to amount of information, type of response, total number of covariates, number of influential covariates, and extent of non-linearity. Measures for comparison are prediction performance, identification of influential covariates, and smoothness of fitted functions. One result is that the mixed model approach returns sparse fits with frequently over-smoothed functions, while the functions are less smooth for the boosting approach and variable selection is less strict. The other approaches are in between with respect to these measures. The boosting procedure is seen to perform very well when little information is available and/or when a large number of covariates is to be investigated. It is somewhat surprising that in scenarios with low information the fitting of a linear model, even with stepwise variable selection, has not much advantage over the fitting of an additive model when the true underlying structure is linear. In cases with more information the prediction performance of all procedures is very similar. So, in difficult data situations the boosting approach can be recommended, in others the procedures can be chosen conditional on the aim of the analysis.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Binder, H., Tutz, G.: Fitting generalized additive models: a comparison of methods. FDM-Preprint 93, University of Freiburg (2006)

  • Breiman, L.: Prediction games and arcing algorithms. Neural Comput. 11, 1493–1517 (1999)

    Article  Google Scholar 

  • Bühlmann, P., Yu, B.: Boosting with the L2 loss: regression and classification. J. Am. Stat. Assoc. 98, 324–339 (2003)

    Article  MATH  Google Scholar 

  • Chambers, J.M., Hastie, T.J.: Statistical Models in S. Wadsworth, Pacific Grove (1992)

    MATH  Google Scholar 

  • Friedman, J.H.: Greedy function approximation: a gradient boosting machine. Ann. Stat. 29, 1189–1232 (2001)

    Article  MATH  Google Scholar 

  • Friedman, J.H., Stuetzle, W.: Projection pursuit regression. J. Am. Stat. Assoc. 76, 817–823 (1981)

    Article  MathSciNet  Google Scholar 

  • Friedman, J., Hastie, T., Rosset, S., Tibshirani, R., Zhu, J.: Statistical behavior and consistency of classification methods based on convex risk minimization: discussion of the paper by T. Zhang. Ann. Stat. 32(1), 102–107 (2004)

    Google Scholar 

  • Green, P.J., Silverman, B.W.: Nonparametric Regression and Generalized Linear Models. Chapman & Hall, London (1994)

    MATH  Google Scholar 

  • Hand, D.J.: Classifier technology and the illusion of progress. Stat. Sci. 21(1), 1–14 (2006)

    Article  MATH  MathSciNet  Google Scholar 

  • Hastie, T., Tibshirani, R.: Generalized additive models. Stat. Sci. 1, 295–318 (1986)

    MathSciNet  Google Scholar 

  • Hastie, T.J., Tibshirani, R.J.: Generalized Additive Models. Chapman & Hall, London (1990)

    MATH  Google Scholar 

  • Hoerl, A.E., Kennard, R.W.: Ridge regression: biased estimation for nonorthogonal problems. Technometrics 12(1), 55–67 (1970)

    Article  MATH  MathSciNet  Google Scholar 

  • Hurvich, C.M., Simonoff, J.S., Tsai, C.: Smoothing parameter selection in nonparametric regression using an improved Akaike information criterion. J. R. Stat. Soc. B 60(2), 271–293 (1998)

    Article  MATH  MathSciNet  Google Scholar 

  • Kim, Y.-J., Gu, C.: Smoothing spline Gaussian regression: more scalable computation via efficient approximation. J. R. Stat. Soc. B 66(2), 337–356 (2004)

    Article  MATH  MathSciNet  Google Scholar 

  • Lee, T.C.M.: Smoothing parameter selection for smoothing splines: a simulation study. Comput. Stat. Data Anal. 42, 139–148 (2003)

    Article  Google Scholar 

  • Lindstrom, M.J.: Penlized estimation of free-knot splines. J. Comput. Graph. Stat. 8(2), 333–352 (1999)

    Article  MathSciNet  Google Scholar 

  • Linton, O.B., Härdle, W.: Estimation of additive regression models with known links. Biometrika 83, 529–540 (1996)

    Article  MATH  MathSciNet  Google Scholar 

  • Marx, B.D., Eilers, P.H.C.: Direct generalized additive modelling with penalized likelihood. Comput. Stat. Data Anal. 28, 193–209 (1998)

    Article  MATH  Google Scholar 

  • McCullagh, P., Nelder, J.A.: Generalized Linear Models, 2nd edn. Chapman & Hall, London (1989)

    Google Scholar 

  • R Development Core Team: R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna (2006). ISBN 3-900051-07-0

    Google Scholar 

  • Ruppert, D.: Selecting the number of knots for penalized splines. J. Comput. Graph. Stat. 11, 735–757 (2002)

    Article  MathSciNet  Google Scholar 

  • Ruppert, D., Wand, M.P., Carroll, R.J.: Semiparametric Regression. Cambridge University Press, Cambridge (2003)

    MATH  Google Scholar 

  • Speed, T.: Comment on “That BLUP is a good thing: the estimation of random effects” by G.K. Robinson. Stat. Sci. 6(1), 42–44 (1991)

    MathSciNet  Google Scholar 

  • Tibshirani, R.: Regression shrinkage and selection via the lasso. J. R. Stat. Soc. B 58(1), 267–288 (1996)

    MATH  MathSciNet  Google Scholar 

  • Tutz, G., Binder, H.: Generalized additive modelling with implicit variable selection by likelihood based boosting. Biometrics 62, 961–971 (2006)

    Article  MATH  MathSciNet  Google Scholar 

  • Wand, M.P.: A comparison of regression spline smoothing procedures. Comput. Stat. 15, 443–462 (2000)

    Article  MATH  MathSciNet  Google Scholar 

  • Wang, Y.: Mixed effects smoothing spline analysis of variance. J. R. Stat. Soc. B 60(1), 159–174 (1998)

    Article  MATH  Google Scholar 

  • Wood, S.N.: Modelling and smoothing parameter estimation with multiple quadratic penalties. J. R. Stat. Soc. B 62(2), 413–428 (2000)

    Article  Google Scholar 

  • Wood, S.N.: Stable and efficient multiple smoothing parameter estimation for generalized additive models. J. Am. Stat. Assoc. 99(467), 673–686 (2004)

    Article  MATH  Google Scholar 

  • Wood, S.N.: Generalized Additive Models. An Introduction with R. Chapman & Hall/CRC, Boca Raton (2006)

    MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Harald Binder.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Binder, H., Tutz, G. A comparison of methods for the fitting of generalized additive models. Stat Comput 18, 87–99 (2008). https://doi.org/10.1007/s11222-007-9040-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11222-007-9040-0

Keywords

Navigation