Abstract
Gold’s original paper on inductive inference introduced a notion of an optimal learner. Intuitively, a learner identifies a class of objects optimally iff there is no other learner that: requires as little of each presentation of each object in the class in order to identify that object, and, for some presentation of some object in the class, requires less of that presentation in order to identify that object. Wiehagen considered this notion in the context of function learning, and characterized an optimal function learner as one that is class-preserving, consistent, and (in a very strong sense) non-U-shaped, with respect to the class of functions learned.
Herein, Gold’s notion is considered in the context of language learning. Intuitively, a language learner identifies a class of languages optimally iff there is no other learner that: requires as little of each text for each language in the class in order to identify that language, and, for some text for some language in the class, requires less of that text in order to identify that language.
Many interesting results concerning optimal language learners are presented. First, it is shown that a characterization analogous to Wiehagen’s does not hold in this setting. Specifically, optimality is not sufficient to guarantee Wiehagen’s conditions; though, those conditions are sufficient to guarantee optimality. Second, it is shown that the failure of this analog is not due to a restriction on algorithmic learning power imposed by non-U-shapedness (in the strong form employed by Wiehagen). That is, non-U-shapedness, even in this strong form, does not restrict algorithmic learning power. Finally, for an arbitrary optimal learner F of a class of languages \(\mathcal {L}\), it is shown that F optimally identifies a subclass \(\mathcal {K}\) of \(\mathcal {L}\) iff F is class-preserving with respect to \(\mathcal {K}\).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Angluin, D.: Inductive inference of formal languages from positive data. Information and Control 45(2), 117–135 (1980)
Barzdin, J.M.: Inductive inference of automata, functions, and programs. American Mathematical Society Translations: Series 2 109, 107–112 (1977); In: Proceeedings of the 20th International Conference of Mathematicians (1974) (appeared originally) (in Russian)
Blum, L., Blum, M.: Toward a mathematical theory of inductive inference. Information and Control 28(2), 125–155 (1975)
Baliga, G., Case, J., Merkle, W., Stephan, F., Wiehagen, W.: When unlearning helps. Information and Computation 206(5), 694–709 (2008)
Bārzdiņš, J., Freivalds, R.: Prediction and limiting synthesis of recursively enumerable classes of functions. Latvijas Valsts Univ. Zinatn. Raksti 210, 101–111 (1974) (in Russian)
Carlucci, L., Case, J., Jain, S., Stephan, F.: Results on memory-limited U-shaped learning. Information and Computation 205(10), 1551–1573 (2007)
Carlucci, L., Case, J., Jain, S., Stephan, F.: Non-U-shaped vacillatory and team learning. Journal of Computer and Systems Sciences 74(4), 409–430 (2008)
Case, J., Moelius, S.E.: Optimal language learning (expanded version). Technical report, University of Delaware (2008), http://www.cis.udel.edu/~moelius/publications
Case, J., Moelius, S.E.: U-shaped, iterative, and iterative-with-counter learning. Machine Learning 72(1-2), 63–88 (2008)
Case, J., Smith, C.: Comparison of identification criteria for machine inductive inference. Theoretical Computer Science 25(2), 193–220 (1983)
Fulk, M.: Prudence and other conditions on formal language learning. Information and Computation 85(1), 1–11 (1990)
Gold, E.M.: Language identification in the limit. Information and Control 10(5), 447–474 (1967)
Jain, S., Osherson, D., Royer, J., Sharma, A.: Systems that Learn: An Introduction to Learning Theory, 2nd edn. MIT Press, Cambridge (1999)
Osherson, D., Stob, M., Weinstein, S.: Systems that Learn: An Introduction to Learning Theory for Cognitive and Computer Scientists, 1st edn. MIT Press, Cambridge (1986)
Pitt, L.: Inductive inference, DFAs, and computational complexity. In: Jantke, K.P. (ed.) AII 1989. LNCS, vol. 397, pp. 18–44. Springer, Heidelberg (1989)
Rogers, H.: Theory of Recursive Functions and Effective Computability. McGraw Hill, New York (1967) (Reprinted, MIT Press, 1987)
Wiehagen, R.: A thesis in inductive inference. In: Dix, J., Schmitt, P.H., Jantke, K.P. (eds.) NIL 1990. LNCS, vol. 543, pp. 184–207. Springer, Heidelberg (1991)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2008 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Case, J., Moelius, S.E. (2008). Optimal Language Learning. In: Freund, Y., Györfi, L., Turán, G., Zeugmann, T. (eds) Algorithmic Learning Theory. ALT 2008. Lecture Notes in Computer Science(), vol 5254. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-87987-9_34
Download citation
DOI: https://doi.org/10.1007/978-3-540-87987-9_34
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-87986-2
Online ISBN: 978-3-540-87987-9
eBook Packages: Computer ScienceComputer Science (R0)