Skip to main content

Learning Intermediate Concepts

Extended Abstract

  • Conference paper
  • First Online:
  • 501 Accesses

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 2225))

Abstract

In most concept learning problems considered so far by the learning theory community, the instances are labeled by a single unknown target. However, in some situations, although the target concept may be quite complex when expressed as a function of the attribute values of the instance, it may have a simple relationship with some intermediate (yet to be learned) concepts. In such cases, it may be advantageous to learn both these intermediate concepts and the target concept in parallel, and use the intermediate concepts to enhance our approximation of the target concept.

In this paper, we consider the problem of learning multiple interrelated concepts simultaneously. To avoid stability problem, we assume that the dependency relations among the concepts are not cyclical and hence can be expressed using a directed acyclic graph (not known to the learner). We investigate this learning problem in various popular theoretical models: mistake bound model, exact learningmo del and probably approximately correct (PAC) model.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. D. Angluin, M. Frazier, and L. Pitt. Learning conjunctions of Horn clauses. Machine Learning, 9:147–164, 1992.

    MATH  Google Scholar 

  2. D. Angluin, L. Hellerstein, and M. Karpinski. Learning read-once formulas with queries. J. ACM, 40:185–210, 1993.

    Article  MathSciNet  Google Scholar 

  3. D. Angluin and M. Krikis. Learningwith malicious membership queries and exceptions. In Proc. 7th Annu. ACM Workshop on Comput. Learning Theory, pages 57–66. ACM Press, New York, NY, 1994.

    Google Scholar 

  4. Dana Angluin, Mārtinš Krikis, Robert H. Sloan, and György Turán. Malicious omissions and errors in answers to membership queries. Machine Learning, 28:211–255, 1997.

    Article  Google Scholar 

  5. D. Angluin. Queries and concept learning. Machine Learning, 2(4):319–342, April 1988.

    MathSciNet  Google Scholar 

  6. D. Angluin and D. K. Slonim. Learning monotone DNF with an incomplete membership oracle. In Proc. 4th Annu. Workshop on Comput. Learning Theory, pages 139–146, San Mateo, CA, 1991. Morgan Kaufmann.

    Google Scholar 

  7. David Barrington. Bounded-width polynomial-size branching programs recognize exactly those language in nc. JCSS, 38:150–164, 1989.

    MathSciNet  MATH  Google Scholar 

  8. Jonathan Baxter. Learningin ternal representations. In Proc. 8th Annu. Conf. on Comput. Learning Theory, pages 311–320. ACM Press, New York, NY, 1995.

    Google Scholar 

  9. Jonathan Baxter. A Bayesian/information theoretic model of learningto learn via multiple task sampling. Machine Learning, 28:7–39, 1997.

    Article  Google Scholar 

  10. Avrim Blum, Prasad Chalasani, Sally A. Goldman, and Donna K. Slonim. Learningwith unreliable boundary queries. In Proc. 8th Annu. Conf. on Comput. Learning Theory, pages 98–107. ACM Press, New York, NY, 1995.

    Google Scholar 

  11. A. Blumer, A. Ehrenfeucht, D. Haussler, and M. K. Warmuth. Learnability and the Vapnik-Chervonenkis dimension. J. ACM, 36(4):929–965, 1989.

    Article  MathSciNet  Google Scholar 

  12. Nader H. Bshouty, Christino Tamon, and David K. Wilson. On learning width two branchingprog rams. In Proc. 9th Annu. Conf. on Comput. Learning Theory, pages 224–227. ACM Press, New York, NY, 1996.

    Google Scholar 

  13. Rich Caruana. Algorithms and applications for multitask learning. In Proc. 13th International Conference on Machine Learning, pages 87–95. Morgan Kaufmann, 1996.

    Google Scholar 

  14. R. Caruana. A dozen tricks with multitask learning. Lecture Notes in Computer Science, 1524:165–187, 1998.

    Google Scholar 

  15. N. Cesa-Bianchi, P. Long, and M. K. Warmuth. Worst-case quadratic loss bounds for on-line prediction of linear functions by gradient descent. IEEE Transactions on Neural Networks, 1995. To appear. An extended abstract appeared in COLT’ 93.

    Google Scholar 

  16. Rich Caruana, Lorien Pratt, and Sebastian Thrun. Multitask learning. Machine Learning, 28:41, 1997.

    Article  Google Scholar 

  17. Thomas G. Dietterich, Hermann Hild, and Ghulum Bakiri. A comparison of ID3 and backpropagation for english text-to-speech mapping. Machine Learning, 18:51–80, 1995.

    Google Scholar 

  18. Ricard Gavaldà and David Guijarro. Learningordered binary decision diagrams. In Proc. 6th Int. Workshop on Algorithmic Learning Theory, pages 228–238. Springer-Verlag, 1995.

    Google Scholar 

  19. D. P. Helmbold, J. Kivinen, and M. K. Warmuth. Worst-case loss bounds for sigmoided linear neurons. In Proc. 1996 Neural Information Processing Conference, 1996. To appear.

    Google Scholar 

  20. M. Kearns. Efficient noise-tolerant learningfrom statistical queries. In Proc. 25th Annu. ACM Sympos. Theory Comput., pages 392–401. ACM Press, New York, NY, 1993.

    Google Scholar 

  21. Michael J. Kearns, Robert E. Schapire, and Linda M. Sellie. Toward efficient agnostic learning. Machine Learning, 17(2/3):115–142, 1994.

    MATH  Google Scholar 

  22. J. Kivinen and M. K. Warmuth. Exponentiated gradient versus gradient descent for linear predictors. Technical Report UCSC-CRL-94-16, University of California, Santa Cruz, Computer Research Laboratory, June 1994. Revised December 7, 1995. An extended abstract to appeared in the STOC 95, pp. 209–218.

    Google Scholar 

  23. N. Littlestone. Learningwhen irrelevant attributes abound: A new linearthreshold algorithm. Machine Learning, 2:285–318, 1988.

    Google Scholar 

  24. Ronald L. Rivest. Learningdecision lists. Machine Learning, 2:229–246, 1987.

    Google Scholar 

  25. Steven C. Suddarth and Alistair D. C. Holden. Symbolic-neural systems and the use of hints for developingcomplex systems. International Journal of Man-Machine Studies, 35(3):291–311, 1991.

    Article  Google Scholar 

  26. S. C. Suddarth and Y. L. Kergosien. Rule-injection hints as a means of improvingnet work performance and learningtime. In Proceedings of the EURASIP Workshop on Neural Networks, pages 120–129, Sesimbra, Portugal, February 1990. EURASIP.

    Google Scholar 

  27. L. G. Valiant. A theory of the learnable. Commun. ACM, 27(11):1134–1142, November 1984.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2001 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Kwek, S.S. (2001). Learning Intermediate Concepts. In: Abe, N., Khardon, R., Zeugmann, T. (eds) Algorithmic Learning Theory. ALT 2001. Lecture Notes in Computer Science(), vol 2225. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45583-3_13

Download citation

  • DOI: https://doi.org/10.1007/3-540-45583-3_13

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-42875-6

  • Online ISBN: 978-3-540-45583-7

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics