Skip to main content

k-Armed Bandit

  • Reference work entry
  • First Online:
Encyclopedia of Machine Learning and Data Mining

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 699.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
USD 949.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Recommended Reading

  • Auer P, Cesa-Bianchi N, Freund Y, Schapire RE (2002) The non-stochastic multi-armed bandit problem. SIAM J Comput 32(1):48–77

    Article  MathSciNet  MATH  Google Scholar 

  • Berry D, Fristedt B (1985) Bandit problems: sequential allocation of experiments. Chapman and Hall, London/New York

    Book  MATH  Google Scholar 

  • Cesa-Bianchi N, Lugosi G (2006) Prediction, learning, and games. Cambridge University Press, New York

    Book  MATH  Google Scholar 

  • Gittins JC (1989) Multi-armed bandit allocation indices. Wiley, New York

    MATH  Google Scholar 

  • Gittins J, Jones D (1972) A dynamic allocation index for sequential design of experiments. In: Progress in statistics, European meeting of statisticians, Budapest, vol 1, pp 241–266

    MATH  Google Scholar 

  • Lai TL, Robbins H (1985) Asymptotically efficient adaptive allocation rules. Adv Appl Math 6:4–22

    Article  MathSciNet  MATH  Google Scholar 

  • Mannor S, Tsitsiklis JN (2004) The sample complexity of exploration in the multi-armed bandit problem. J Mach Learn Res 5:623–648

    MathSciNet  MATH  Google Scholar 

  • Robbins H (1952) Some aspects of the sequential design of experiments. Bull Am Math Soc 55:527–535

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer Science+Business Media New York

About this entry

Cite this entry

Mannor, S. (2017). k-Armed Bandit. In: Sammut, C., Webb, G.I. (eds) Encyclopedia of Machine Learning and Data Mining. Springer, Boston, MA. https://doi.org/10.1007/978-1-4899-7687-1_424

Download citation

Publish with us

Policies and ethics