skip to main content
10.1145/1553374.1553390acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlConference Proceedingsconference-collections
research-article

Robust bounds for classification via selective sampling

Published:14 June 2009Publication History

ABSTRACT

We introduce a new algorithm for binary classification in the selective sampling protocol. Our algorithm uses Regularized Least Squares (RLS) as base classifier, and for this reason it can be efficiently run in any RKHS. Unlike previous margin-based semi-supervised algorithms, our sampling condition hinges on a simultaneous upper bound on bias and variance of the RLS estimate under a simple linear label noise model. This fact allows us to prove performance bounds that hold for an arbitrary sequence of instances. In particular, we show that our sampling strategy approximates the margin of the Bayes optimal classifier to any desired accuracy ε by asking Õ (d2) queries (in the RKHS case d is replaced by a suitable spectral quantity). While these are the standard rates in the fully supervised i.i.d. case, the best previously known result in our harder setting was Õ (d34). Preliminary experiments show that some of our algorithms also exhibit a good practical performance.

References

  1. Auer, P. (2002). Using confidence bounds for exploitation-exploration trade-offs. Journal of Machine Learning Research, 3, 397--422. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Balcan, M., Beygelzimer, A., & Langford, J. (2006). Agnostic active learning. Proc. of the 23rd International Conference on Machine Learning (pp. 65--72). Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Balcan, M., Broder, A., & Zhang, T. (2007). Margin-based active learning. Proceedings of the 20th Annual Conference on Learning Theory (pp. 35--50). Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Cavallanti, G., Cesa-Bianchi, N., & Gentile, C. (2007). Tracking the best hyperplane with a simple budget Perceptron. Machine Learning, 69, 143--167. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Cavallanti, G., Cesa-Bianchi, N., & Gentile, C. (2009). Linear classification and selective sampling under low noise conditions. In Advances in Neural Information Processing Systems 21 (pp. 249--256).Google ScholarGoogle Scholar
  6. Cesa-Bianchi, N., Gentile, C., & Zaniboni, L. (2006a). Incremental algorithms for hierarchical classification. Journal of Machine Learning Research, 7, 31--54. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Cesa-Bianchi, N., Gentile, C., & Zaniboni, L. (2006b). Worst-case analysis of selective sampling for linear classification. Journal of Machine Learning Research, 7, 1025--1230. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Cohn, R., Atlas, L., & Ladner, R. (1990). Training connectionist networks with queries and selective sampling. In Advances in Neural Information Processing Systems 2 (pp. 566--573). Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Dasgupta, S., Hsu, D., & Monteleoni, C. (2008). A general agnostic active learning algorithm. In Advances in Neural Information Processing Systems 21 (pp. 353--360).Google ScholarGoogle Scholar
  10. Dasgupta, S., Kalai, A. T., & Monteleoni, C. (2005). Analysis of perceptron-based active learning. Proceedings of the 18th Annual Conference on Learning Theory (pp. 249--263). Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Dekel, O., Shalev-Shwartz, S., & Singer, Y. (2007). The Forgetron: A kernel-based Perceptron on a budget. SIAM Journal on Computing, 37, 1342--1372. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Freund, Y., Seung, S., Shamir, E., & Tishby, N. (1997). Selective sampling using the query by committee algorithm. Machine Learning, 28, 133--168. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Li, L., Littman, M., & Walsh, T. (2008). Knows what it knows: a framework for self-aware learning. Proceedings of the 25th International Conference on Machine Learning (pp. 568--575). Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Orabona, F., Keshet, J., & Caputo, B. (2008). The Projectron: a bounded kernel-based Perceptron. Proceedings of the 25th International Conference on Machine Learning (pp. 720--727). Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Strehl, A., & Littman, M. (2008). Online linear regression and its application to model-based reinforcement learning. In Advances in Neural Information Processing Systems 20 (pp. 631--638).Google ScholarGoogle Scholar
  16. Weston, J., Bordes, A., & Bottou, L. (2005). Online (and offline) on an even tighter budget. Proceedings of the 10th International Workshop on Artificial Intelligence and Statistics (pp. 413--420).Google ScholarGoogle Scholar

Index Terms

  1. Robust bounds for classification via selective sampling

                    Recommendations

                    Comments

                    Login options

                    Check if you have access through your login credentials or your institution to get full access on this article.

                    Sign in
                    • Published in

                      cover image ACM Other conferences
                      ICML '09: Proceedings of the 26th Annual International Conference on Machine Learning
                      June 2009
                      1331 pages
                      ISBN:9781605585161
                      DOI:10.1145/1553374

                      Copyright © 2009 Copyright 2009 by the author(s)/owner(s).

                      Publisher

                      Association for Computing Machinery

                      New York, NY, United States

                      Publication History

                      • Published: 14 June 2009

                      Permissions

                      Request permissions about this article.

                      Request Permissions

                      Check for updates

                      Qualifiers

                      • research-article

                      Acceptance Rates

                      Overall Acceptance Rate140of548submissions,26%

                    PDF Format

                    View or Download as a PDF file.

                    PDF

                    eReader

                    View online with eReader.

                    eReader