Skip to main content

On Variants of Iterative Learning

  • Conference paper
  • First Online:
Discovey Science (DS 1998)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 1532))

Included in the following conference series:

  • 580 Accesses

Abstract

Within the present paper, we investigate the principal learning capabilities of iterative learners in some more details. The general scenario of iterative learning is as follows. An iterative learner successively takes as input one element of a text (an informant) of a target concept as well as its previously made hypothesis, and outputs a new hypothesis about the target concept. The sequence of hypotheses has to converge to a hypothesis correctly describing the target concept

We study the following variants of this basic scenario. First, we consider the case that an iterative learner has to learn on redundant texts or informants, only. A text (an informant) is redundant, if it contains every data item infinitely many times. This approach guarantees that relevant information is, in principle, accessible at any time in the learning process. Second, we study a version of iterative learning, where an iterative learner is supposed to learn independent on the choice of the initial hypothesis. In contrast, in the basic scenario of iterative inference, it is assumed that the initial hypothesis is the same for every learning task which allows certain coding tricks

We compare the learning capabilities of all models of iterative learning from text and informant, respectively, to one another as well as to finite inference, conservative identification, and learning in the limit from text and informant, respectively

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Angluin, D., Inductive inference of formal languages from positive data, Information and Control 45, 117–135, 1980.

    Article  MATH  MathSciNet  Google Scholar 

  2. Blum, M., A machine independent theory of the complexity of recursive functions, Journal of the ACM 14, 322–336, 1967.

    Article  MATH  MathSciNet  Google Scholar 

  3. Gold, E.M., Language identification in the limit, Information and Control 10, 447–474, 1967.

    Article  MATH  Google Scholar 

  4. Hopcroft, J.E., and Ullman, J.D., “Formal Languages and their Relation to Automata”, Addison-Wesley, 1969.

    Google Scholar 

  5. Kinber, E., and Stephan, F., Language learning from texts: Mind changes, limited memory and monotonicity, in “Proceedings 8th Annual ACM Conference on Computational Learning Theory,” pp. 182–189, ACM Press, 1995.

    Google Scholar 

  6. Kolodner, J.K., An introduction to case-based reasoning. Artificial Intelligence Review 6, 3–34, 1992.

    Article  Google Scholar 

  7. Lange, S., and Zeugmann, T., Types of monotonicl anguage learning and their characterization, in “Proceedings 5th Annual ACM Workshop on Computational Learning Theory,” pp. 377–390, ACM Press, 1992.

    Google Scholar 

  8. Lange, S., and Zeugmann, T., Language learning in dependence on the space of hypotheses, in “Proceedings 6th Annual ACM Conference on Computational Learning Theory,” pp. 127–136, ACM Press, 1993.

    Google Scholar 

  9. Lange, S., and Zeugmann, T., Incremental learning from positive data, Journal of Computer and System Sciences 53, 88–103, 1996.

    Article  MATH  MathSciNet  Google Scholar 

  10. Lange, S., and Zeugmann, T., Set-driven and rearrangement-independent learning of recursive languages, Mathematical Systems Theory 29, 599–634, 1996.

    MATH  MathSciNet  Google Scholar 

  11. Rivest, R., Learning decision lists, Machine Learning 2, 229–246, 1987.

    MathSciNet  Google Scholar 

  12. Valiant, L.G., A theory of the learnable, Communications of the ACM 27, 1134–1142, 1984.

    Article  MATH  Google Scholar 

  13. Wexler, K., and Culicover, P., “Formal Principles of Language Acquisition”, MIT Press, 1980.

    Google Scholar 

  14. Wiehagen, R., Limes-Erkennung rekursiver Funktionen durch spezielle Strategien, Journal of Information Processing and Cybernetics (EIK) 12, 93–99, 1976.

    Google Scholar 

  15. Zeugmann, T., and Lange, S., A guided tour across the boundaries of learning recursive languages, in “AlgorithmicL earning for Knowledge-Based Systems,” Lecture Notes in Artificial Intelligence, Vol. 961, pp. 193–262, Springer-Verlag, 1995.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1998 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Lange, S., Grieser, G. (1998). On Variants of Iterative Learning. In: Arikawa, S., Motoda, H. (eds) Discovey Science. DS 1998. Lecture Notes in Computer Science(), vol 1532. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-49292-5_7

Download citation

  • DOI: https://doi.org/10.1007/3-540-49292-5_7

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-65390-5

  • Online ISBN: 978-3-540-49292-4

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics