Abstract
The paper presents the system LERS for rule induction. The system handles inconsistencies in the input data due to its usage of rough set theory principle. Rough set theory is especially well suited to deal with inconsistencies. In this approach, inconsistencies are not corrected. Instead, system LERS computes lower and upper approximations of each concept. Then it induces certain rules and possible rules. The user has the choice to use the machine learning approach or the knowledge acquisition approach. In the first case, the system induces a single minimal discriminant description for each concept. In the second case, the system induces all rules, each in the minimal form, that can be induced from the input data. In both cases, the user has a choice between the local or global approach.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Aha, D. W. and D. Kibler. Noise-tolerant instance-based learning algorithms. Proc. of the IJCAI 89, 11th Int. Joint Conf. on Al,794–799.
Bergadano, F., A. Giordana, and L. Saitta. Automated concept acquisition in noisy environment. IEEE Trans. PAMI 10, 1988, 555–578.
Bergadano, F. and L. Saitta. On the error probability of Boolean concept description. Proc. of the 4th European Working Session on Learning, 1989, 25–36.
Berzuini, C. Combining symbolic learning techniques and statistical regression analysis. Proc. of the AAAI-88, 7th Nat. Conf. on AI,612–617.
Berzuini, C. Partially supervised learning from examples with the aid of statistical regression analysis. Proc of the ISMIS-88, 3rd Int. Symp. on Methodologies for Intelligent Systems,281–292.
Boucheron, S. and J. Sallantin. Leamability in the presence of noise. Proc. of the 3rd European Working Session on Learning, 1988, 25–35.
Buntine, W. Decision tree induction systems: A Bayesian analysis. In Uncertainty in AI 3, L. N. Kanal, T. S. Lewitt, and J. L. Lemmer (eds.), Elsevier, 1989, 109–127.
Buntine, W. Learning classification rules using Bayes. Proc. of the 6th Int. Workshop on Machine Learning, 1989, 94–98.
Buntine, W. A critique of the Valiant model. Proc. of the IJCAI-89, 11th Int. Joint Conf. on AI,837–842.
Breiman, L., J. H. Friedman, R. A. Olshen, and C. J. Stone. Classification and Regression Trees. Wadsworth & Brooks, 1984.
Budihardjo, A., J. W. Grzymala-Busse, and L. Woolery. Program LERS_LB 2.5 as a tool for knowledge acquisition in nursing. Proc. of the 4th Mt. Conf. on Industrial & Engineering Applications of Artificial Intelligence & Expert Systems, 1991, 735–740.
Chan, C. C. and J. W. Grzymala-Busse. Rough-set boundaries as a tool for learning rules from examples. Proc. of the ISMIS-89, 4th Int. Symp. on Methodologies for Intelligent Systems,281–288.
Chan, C. C. and J. W. Grzymala-Busse. On the attribute redundancy and the learning programs ID3, PRISM, and LEM. Submitted for publication
Chan, P. K. Inductive learning with BCT. Proc. of the 6th lnt. Workshop on Machine Learning, 1989, 104–108.
] Chang, K. C. C. and A. K. C. Wong. Performance analysis of a probabilistic inductive learning system. Proc. of the 7th Conf. on Machine Learning, 1990, 16–23.
Cheeseman, P., J. Kelly, M. Self, J. Stutz, W. Taylor, and D. Freeman. AutoClass• A Bayesian classification system. Proc. of the 5th Int. Conf. on Machine Learning, 1988, 54–64.
Cheeseman, P., M. Self, J. Kelly, J. Stutz, W. Taylor, and D. Freeman. Bayesian classification. Proc. of the AAAI-88, 7th Nat. Conf. on AI,607–610.
Clark, P. and T. Niblett. The CN2 induction algorithm. Machine Learning 3, 1989, 261–283.
Dean, J. S. and J. W. Grzymala-Busse. An overview of the learning from examples module LEM1. Report TR-88–2, Department of Computer Science, University of Kansas, 1988.
Dubois, D. and H. Prade. Epistemic entrenchment and possibilistic logic. Artificial Intelligence 50, 1991, 223–239.
Fang, R. M. and S. L. Crawford. CONSTRUCTOR: A system for the induction of probabilistic models. Proc. of the AAAI-90, 8th Nat. Conf. on AI,762–769.
Goodman, R. M. and P. Smyth. The induction of probabilistic rule sets—the ITRULE algorithm. Proc. of the 6th Int. Workshop on Machine Learning, 1989, 129–132.
Grzymala-Busse, J. W. Knowledge acquisition under uncertainty—A rough set approach. J. Intelligent & Robotic Systems 1, 1988, 3–16.
Grzymala-Busse, J. W. An overview of the LERS1 learning system. Proc. of the 2nd Int. Conf. on Industrial and Engineering Applications of Artificial Intelligence and Expert Systems, 1989, 838–844.
Grzymala-Busse, J. W. On the unknown attribute values in learning from examples. Proc. of the ISMIS’91, 6th Int. Symp. on Methodologies for Intelligent Systems,368377.
Grzymala-Busse, J. W. Managing Uncertainty in Expert Systems. Kluwer Academic Publishers, 1991.
Grzymala-Busse, J. W. and D. J. Sikora. LERS1—A system for learning from examples based on rough sets. Report TR-88–5, Department of Computer Science, University of Kansas, 1988.
Hanson, S. J. and M. Bauer. Machine learning, clustering and polymorphy. In Uncertainty in Artificial Intelligence, L. N. Kanal and J. F. Lemmer (eds.) Elsevier Science Publ., 1986, 415–428.
Haussier, D. Bias, version spaces and Valiant’s learning framework. Proc. of the 4th Int. Workshop on Machine Learning, 1987, 324–336.
Haussier, D. Probably approximately correct learning. Proc. of the AAAI-90, 8th Nat. Conf. on AI,1101–1108.
Hirsch, H. Learning from data with bounded inconsistency. Proc. of the 7th Conf. on Machine Learning, 1990, 32–39.
Kodratoff, Y., M. Manago, and J. Blythe. Generalization and noise. Int. J. Man-Machine Studies 27, 1987, 181–204.
Kononenko, I. ID3, sequential Bayes, naive Bayes and Bayesian neural networks. Proc. of the 4th European Working Session on Learning, 1989, 91–98.
Kononenko, I. and I. Bratko. Information-based evaluation criterion for classifiers performance. Machine Learning 6, 1991, 67–80.
Kwok, S. W. and C. Carter. Multiple decision trees. In Uncertainly in AI 4, R. D. Shachter, T. S. Levitt, L. N. Kanal, J. F. Lemmer (eds.), Elsevier, 1990, 327–335.
Laird, P. D. Learning from Good and Bad Data. Kluwer Academic Publishers, 1988.
Lee, W. D. and S. R. Ray. Rule refinement using the probabilistic rule generator. Proc. of the AAAI-86, 5th Nat. Conf. on AI,442–447.
Lesmo, L., L. Saitta, and P. Torasso. Learning of fuzzy production rules for medical diagnosis. In Approximate reasoning in Decision Analysis, M. M. Gupta and E. Sanchez (eds.), North Holland Publ. Co., 1982, 249–260.
Manago, M. V. and Y. Kodratoff. Noise and knowledge acquisition. Proc. of the IJCAI 87, Int. Joint Conf. on AI,348–354.
Michalski, R. S. A theory and methodology of inductive learning. In Machine Learning, R. S. Michalski, J. G. Carbonell, T. M. Mitchell (eds.), Morgan Kaufmann 1983, 83–134.
Michalski, R. S., I. Mozetic, J. Hong, and N. Lavrac. The AQ15 inductive learning system: An overview and experiments. Report 1260, Department of Computer Science, University of Illinois at Urbana-Champaign, 1986.
Michalski, R. S. How to learn imprecise concepts: A method for employing a two-tiered knowledge representation in learning. Proc of the 4th Int. Workshop on Machine Learning, 1987, 50–58.
Mingers, J. An empirical comparison of selection measures for decision-tree induction. Machine Learning 3, 1989, 319–342.
] Mingers, J. An empirical comparison of pruning methods for decision tree induction. Machine Learning 4, 1989, 227–243.
Nakakuki, Y., Y. Koseki, and M. Tanaka. Inductive learning in probabilistic domain. Proc. of the AAAI-90, 8th Nat. Conf. on Al, 809–814.
Niblett, T. and I. Bratko. Learning decision rules in noisy domains. Proc. of Expert Systems ‘86, the 6th Annual Techn. Conference of the British Computer Society, Specialist Group on Expert Systems, 1986, 25–34.
Pawlak, Z. Rough sets. Int. J. Computer and Information Sci., 11, 1982, 341–356.
Pawlak, Z. Rough Classification. Int. J. Man-Machine Studies 20, 1984, 469–483.
Plaza, E. and R. Lopez de Mantaras. A case-based apprentice that learns from fuzzy examples. Proc. of the 5th Int. Symp. on Methodologies for Intelligent Systems, 1990, 420–427.
Quinlan, J. R. Learning efficient classification procedures and their application to chess end games. In Machine Learning, R. S. Michalski, J. G. Carbonell, T. M. Mitchell (eds.), Morgan Kaufmann Publishers, Inc., 1983, 461–482.
Quinlan, J. R. Induction of decision trees. Machine Learning 1, 1986, 81–106.
Quinlan, J. R. Decision trees as probabilistic classifiers. Proc of the 4th Int. Workshop on Machine Learning, 1987, 31–37.
Quinlan, J. R. Generating production rules from decision trees. Proc. of the 10th Int. Joint Conf. on AI, 1987, 304–307.
Quinlan, J. R. Unknown attribute values in induction. Proc. of the 6th Int. Workshop on Machine Learning, 1989, 164–168.
Quinlan, J. R. The effect of noise on concept learning. In Machine Learning. An Artificial Intelligence Approach. Vol. II, R. S. Michalski, J. G. Carbonell, T. M. Mitchell, (eds). Morgan Kaufmann Publishers, Inc., 1986, 149–166.
Quinlan, J. R. Probabilistic decision trees. In Machine Learning. An Artificial Intelligence Approach. Vol. III, Morgan Kaufmann Publishers, Inc., 1990, 140–152.
Rendell, L. Induction, of and by probability. In Uncertainty in Artificial Intelligence, L. N. Kanal and J. F. Lemmer (eds.), Elsevier Science Publ., 1986, 415–443.
Rivest, R. L. Learning decision lists. Machine Learning 2, 1987, 229–246.
Shackelford, G. and D. Volper. Learning k-DNF with noise in the attributes. Proc. of the 1988 Workshop on Computational Learning Theory,97–101.
Schlimmer, J. C. and D. Fisher. A case study of incremental concept induction. Proc of the AAAI-86, 5th Nat. Conf. on AI,496–501.
Sloan, R. Types of noise in data for concept learning. Proc. of the 1988 Workshop on Computational Learning Theory,91–96.
Spangler, S., U. M. Fayyad, and R. Uthurusamy. Induction of decision trees from inconclusive data. Proc. of the 6th Int. Workshop on Machine Learning,1989, 146150.
Tan, M. and L. Eshelman. Using weighted networks to represent classification knowledge in noisy domains. Proc. of the 5th Int. Conf. on Machine Learning, 1988, 121–134.
Tan, M. and L. Eshelman. The impact of noise on learning. Rep. CMU-CS-88–144, Department of Computer Science, Carnegie Mellon, 1988.
Utgoff, P. E. IDS: An incremental ID3. Proc. of the 5th Int. Conf. on Machine Learning, 1988, 107–120.
Utgoff, P. E. and C. E. Brodley. An incremental method for finding multivariate splits for decision trees. Proc. of the 7th Conf. on Machine Learning, 1990, 58–65.
Wong, S. K. M. and W. Ziarko. On learning and evaluation of decision rules in the context of rough sets. Proc. of the ACM SIGART Int. Symp. on Methodologies for Intelligent Systems, 1986, 308–324.
Valiant, L. G. A theory of learnable. Corn. ACM 27, 1984, 1134–1142.
Valiant, L. G. Learning disjunction of conjunctions. Proc. of the 9th Int. Joint Conf. on AI, 1985, 560–566.
Yasdi, R. and W. Ziarko. An expert system for conceptual schema design: A machine learning approach. hit. J. Man-Machine Studies 29, 1988, 351–376.
Ziarko, W. and J. D. Katzberg. Control algorithm acquisition, analysis, and reduction. In Knowledge-Based System Diagnosis, Supervision and Control, S. T. Tzafestas (ed.), Plenum Press, 1989, 167–178.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1992 Springer Science+Business Media Dordrecht
About this chapter
Cite this chapter
Grzymala-Busse, J.W. (1992). LERS-A System for Learning from Examples Based on Rough Sets. In: Słowiński, R. (eds) Intelligent Decision Support. Theory and Decision Library, vol 11. Springer, Dordrecht. https://doi.org/10.1007/978-94-015-7975-9_1
Download citation
DOI: https://doi.org/10.1007/978-94-015-7975-9_1
Publisher Name: Springer, Dordrecht
Print ISBN: 978-90-481-4194-4
Online ISBN: 978-94-015-7975-9
eBook Packages: Springer Book Archive