Abstract
In this article we present the new machine learning framework called NRough. It is focused on rough set based algorithms for feature selection and classification i.e. computation of various types of decision reducts, bireducts, decision reduct ensembles and rough set inspired decision rule induction. Moreover, the framework contains other routines and algorithms for supervised and unsupervised learning. NRough is written in C# and compliant with .NET Common Language Specification (CLS). Its architecture allows easy extendability and integration.
References
Pawlak, Z., Skowron, A.: Rudiments of rough sets. Inf. Sci. 177(1), 3–27 (2007)
Widz, S., Ślęzak, D.: Rough set based decision support - models easy to interpret. In: Peters, G., Lingras, P., Ślęzak, D., Yao, Y. (eds.) Rough Sets: Selected Methods and Applications in Management and Engineering, pp. 95–112. Springer, Heidelberg (2012)
Agrawal, R., Imielinski, T., Swami, A.N.: Mining association rules between sets of items in large databases. In: Buneman, P., Jajodia, S. (eds.) Proceedings of the 1993 ACM SIGMOD International Conference on Management of Data, Washington, D.C., 26–28 May 1993, pp. 207–216. ACM Press (1993)
Rokach, L., Maimon, O.: Data Mining with Decision Trees: Theory and Applications. World Scientific, Singapore (2014)
Świniarski, R.W., Skowron, A.: Rough set methods in feature selection and recognition. Pattern Recogn. Lett. 24(6), 833–849 (2003)
Nguyen, H.S., Ślęzak, D.: Approximate reducts and association rules - correspondence and complexity results. In: Zhong, N., Skowron, A., Ohsuga, S. (eds.) RSFDGrC 1999. LNCS, vol. 1711, pp. 137–145. Springer, Heidelberg (1999). doi:10.1007/978-3-540-48061-7_18
Widz, S.: NRough framework git repository (2017). https://www.github.org/nrough/
Widz, S.: NRough framework website (2017). http://www.nrough.net
Bazan, J.G., Szczuka, M.S.: The rough set exploration system. In: Peters, J.F., Skowron, A. (eds.) Transactions on Rough Sets III. LNCS, vol. 3400, pp. 37–56. Springer, Heidelberg (2005)
Lichman, M.: UCI machine learning repository (2013)
Ślęzak, D.: Rough sets and functional dependencies in data: foundations of association reducts. In: Gavrilova, M.L., Tan, C.J.K., Wang, Y., Chan, K.C.C. (eds.) Transactions on Computational Science V. LNCS, vol. 5540, pp. 182–205. Springer, Heidelberg (2009). doi:10.1007/978-3-642-02097-1_10
Ślęzak, D., Ziarko, W.: The investigation of the Bayesian rough set model. Int. J. Approximate Reason. 40(1–2), 81–91 (2005)
Widz, S., Ślęzak, D.: Attribute subset quality functions over a universe of weighted objects. In: Kryszkiewicz, M., Cornelis, C., Ciucci, D., Medina-Moreno, J., Motoda, H., Raś, Z.W. (eds.) RSEISP 2014. LNCS, vol. 8537, pp. 99–110. Springer, Cham (2014). doi:10.1007/978-3-319-08729-0_9
Stawicki, S., Ślęzak, D., Janusz, A., Widz, S.: Decision bireducts and decision reducts - a comparison. Int. J. Approximate Reason. 84, 75–109 (2017)
Stawicki, S., Widz, S.: Decision bireducts and approximate decision reducts: comparison of two approaches to attribute subset ensemble construction. In: 2012 Federated Conference on Computer Science and Information Systems (FedCSIS), pp. 331–338. IEEE (2012)
Widz, S., Stawicki, S.: Generalized majority decision reducts. In: 2016 Federated Conference on Computer Science and Information Systems (FedCSIS), pp. 165–174. IEEE (2016)
Kuncheva, L.I., Diez, J.J.R., Plumpton, C.O., Linden, D.E.J., Johnston, S.J.: Random subspace ensembles for fMRI classification. IEEE Trans. Med. Imaging 29(2), 531–542 (2010)
Breiman, L.: Bagging predictors. Mach. Learn. 24(2), 123–140 (1996)
Kuncheva, L.I.: Combining Pattern Classifiers: Methods and Algorithms. Wiley, Hoboken (2014)
Freund, Y., Schapire, R.E.: Experiments with a new boosting algorithm. In: Saitta, L. (ed.) ICML, pp. 148–156. Morgan Kaufmann, Burlington (1996)
Widz, S.: Boosting approximate reducts. In: Techniki informacyjne: teoria i zastosowania: wybrane problemy, Instytut Badań Systemowych PAN, vol. 5, no. 17, pp. 129–148 (2015)
Ślęzak, D., Widz, S.: Is it important which rough-set-based classifier extraction and voting criteria are applied together? In: Szczuka, M., Kryszkiewicz, M., Ramanna, S., Jensen, R., Hu, Q. (eds.) RSCTC 2010. LNCS, vol. 6086, pp. 187–196. Springer, Heidelberg (2010). doi:10.1007/978-3-642-13529-3_21
Kohavi, R.: The power of decision tables. In: Lavrac, N., Wrobel, S. (eds.) ECML 1995. LNCS, vol. 912, pp. 174–189. Springer, Heidelberg (1995). doi:10.1007/3-540-59286-5_57
Quinlan, J.R.: C4.5: Programs for Machine Learning. Elsevier, Amsterdam (2014)
Quinlan, J.R.: Induction of decision trees. Mach. Learn. 1(1), 81–106 (1986)
Quinlan, J.R.: Simplifying decision trees. Int. J. Man-Mach. Stud. 27(3), 221–234 (1987)
Holte, R.C.: Very simple classification rules perform well on most commonly used datasets. Mach. Learn. 11(1), 63–90 (1993)
Gurrutxaga, I., Arbelaitz, O., Martín, J.I., Muguerza, J., Pérez, J.M., Perona, I.: SIHC: a stable incremental hierarchical clustering algorithm. In: ICEIS, vol. 2, pp. 300–304 (2009)
Fayyad, U.M., Irani, K.B.: Multi-interval discretization of continuous-valued attributes for classification learning. In: Thirteenth International Joint Conference on Artificial Intelligence, vol. 2, pp. 1022–1027. Morgan Kaufmann Publishers (1993)
Kononenko, I.: On biases in estimating multi-valued attributes. In: 14th International Joint Conference on Articial Intelligence, pp. 1034–1040 (1995)
Friedman, J., Hastie, T., Tibshirani, R.: The Elements of Statistical Learning: Data Mining, Inference, and Prediction. Springer Series in Statistics. Springer, Heidelberg (2009)
R Development Core Team: R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria (2008). ISBN 3-900051-07-0
Rüegg, C., Marcus, C.: Math.NET numerics (2017). https://numerics.mathdotnet.com/. Accessed 11 Feb 2017
Poole, C., Prouse, R., Busoli, S., Colvin, N.: NUnit framework (2017) https://www.nunit.org/. Accessed 11 Feb 2017
Perraud, J.M.: R.NET github repository (2017). https://github.com/jmp. 75/rdotnet. Accessed 11 Feb 2017
Free Software Foundation: Gnu lesser general public license. https://www.gnu.org/licenses/lgpl-3.0.en.html. Accessed 11 Feb 2017
Grzymala-Busse, J.W.: LERS-a data mining system. In: Maimon, O., Rokach, L. (eds.) Data Mining and Knowledge Discovery Handbook, pp. 1347–1351. Springer, Heidelberg (2005)
Komorowski, J., Øhrn, A., Skowron, A.: Case studies: public domain, multiple mining tasks systems: ROSETTA rough sets. In: Handbook of Data Mining and Knowledge Discovery, pp. 554–559. Oxford University Press, Inc. (2002)
Predki, B., Słowiński, R., Stefanowski, J., Susmaga, R., Wilk, S.: ROSE - software implementation of the rough set theory. In: Polkowski, L., Skowron, A. (eds.) RSCTC 1998. LNCS, vol. 1424, pp. 605–608. Springer, Heidelberg (1998). doi:10.1007/3-540-69115-4_85
Wojna, A.: RSESLib. (2017). http://rseslib.mimuw.edu.pl/. Accessed 11 Feb 2017
Riza, L.S., Janusz, A., Bergmeir, C., Cornelis, C., Herrera, F., Ślęzak, D., Benítez, J.M., et al.: Implementing algorithms of rough set theory and fuzzy rough set theory in the R package “roughsets”. Inf. Sci. 287, 68–89 (2014)
Janusz, A., Stawicki, S., Szczuka, M., Ślęzak, D.: Rough set tools for practical data exploration. In: Ciucci, D., Wang, G., Mitra, S., Wu, W.-Z. (eds.) RSKT 2015. LNCS, vol. 9436, pp. 77–86. Springer, Cham (2015). doi:10.1007/978-3-319-25754-9_7
Barga, R., Fontama, V., Tok, W.H., Cabrera-Cordon, L.: Predictive Analytics with Microsoft Azure Machine Learning. Springer, Heidelberg (2015)
Souza, C.R.: The Accord.NET framework, SãoCarlos, Brazil (2014). http://accord-framework.net. Accessed 11 Feb 2017
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendix NRough Code Samples
Appendix NRough Code Samples
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Widz, S. (2017). Introducing NRough Framework. In: Polkowski, L., et al. Rough Sets. IJCRS 2017. Lecture Notes in Computer Science(), vol 10313. Springer, Cham. https://doi.org/10.1007/978-3-319-60837-2_53
Download citation
DOI: https://doi.org/10.1007/978-3-319-60837-2_53
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-60836-5
Online ISBN: 978-3-319-60837-2
eBook Packages: Computer ScienceComputer Science (R0)