Skip to main content

Bias-Variance Trade-Offs: Novel Applications

  • Reference work entry
  • First Online:
Encyclopedia of Machine Learning and Data Mining
  • 214 Accesses

Definition

Consider a given random variable \(\underline{F}\) and a random variable that we can modify, \(\hat{\underline{F}}\). We wish to use a sample of \(\hat{\underline{F}}\) as an estimate of a sample of F. The mean squared error (MSE) between such a pair of samples is a sum of four terms. The first term reflects the statistical coupling between F and \(\hat{\underline{F}}\) and is conventionally ignored in bias-variance analysis. The second term reflects the inherent noise in F and is independent of the estimator \(\hat{\underline{F}}\). Accordingly, we cannot affect this term. In contrast, the third and fourth terms depend on \(\hat{\underline{F}}\). The third term, called the bias, is independent of the precise samples of both F and \(\hat{\underline{F}}\) and reflects the difference between the means of F and \(\hat{\underline{F}}\). The fourth term, called the variance, is independent of the precise sample of Fand reflects the inherent noise in the estimator as one samples...

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 699.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
USD 949.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Recommended Reading

  • Angluin D (1992) Computational learning theory: survey and selected bibliography. In: Proceedings of the twenty-fourth annual ACM symposium on theory of computing, Victoria. ACM, New York

    Book  Google Scholar 

  • Berger JO (1985) Statistical decision theory and bayesian analysis. Springer, New York

    Book  MATH  Google Scholar 

  • Breiman L (1996a) Bagging predictors. Mach Learn 24(2):123–140

    MATH  Google Scholar 

  • Breiman L (1996b) Stacked regression. Mach Learn 24(1):49–64

    MATH  Google Scholar 

  • Buntine W, Weigend A (1991) Bayesian back-propagation. Complex Syst 5:603–643

    MATH  Google Scholar 

  • Ermoliev YM, Norkin VI (1998) Monte carlo optimization and path dependent nonstationary laws of large numbers. Technical Report IR-98-009. International Institute for Applied Systems Analysis, Austria

    Google Scholar 

  • Lepage GP (1978) A new algorithm for adaptive multidimensional integration. J Comput Phys 27:192–203

    Article  MATH  Google Scholar 

  • Mackay D (2003) Information theory, inference, and learning algorithms. Cambridge University Press, Cambridge

    MATH  Google Scholar 

  • Robert CP, Casella G (2004) Monte Carlo statistical methods. Springer, New York

    Book  MATH  Google Scholar 

  • Rubinstein R, Kroese D (2004) The cross-entropy method. Springer, New York

    Book  MATH  Google Scholar 

  • Smyth P, Wolpert D (1999) Linearly combining density estimators via stacking. Mach Learn 36(1–2):59–83

    Article  Google Scholar 

  • Vapnik VN (1982) Estimation of dependences based on empirical data. Springer, New York

    MATH  Google Scholar 

  • Vapnik VN (1995) The nature of statistical learning theory. Springer, New York

    Book  MATH  Google Scholar 

  • Wolpert DH (1997) On bias plus variance. Neural Comput 9:1211–1244

    Article  Google Scholar 

  • Wolpert DH, Rajnarayan D (2007) Parametric learning and monte carlo optimization. arXiv:0704.1274v1 [cs.LG]

    Google Scholar 

  • Wolpert DH, Strauss CEM, Rajnarayan D (2006) Advances in distributed optimization using probability collectives. Adv Complex Syst 9(4):383–436

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Dev Rajnarayan .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer Science+Business Media New York

About this entry

Cite this entry

Rajnarayan, D., Wolpert, D. (2017). Bias-Variance Trade-Offs: Novel Applications. In: Sammut, C., Webb, G.I. (eds) Encyclopedia of Machine Learning and Data Mining. Springer, Boston, MA. https://doi.org/10.1007/978-1-4899-7687-1_28

Download citation

Publish with us

Policies and ethics