Abstract
We consider a problem of selection of parameters in a classifier based on the average of kernel density estimators where each estimator corresponds to a different data “resolution”. The selection is based on adjusting parameters of the estimators to minimize a substitute of the misclassification ratio. We experimentally compare the misclassification ratio and parameters selected for benchmark data sets by the introduced algorithm with these values of the algorithm’s baseline version. In order to place the classification results in a wider context, we compare them with results of other popular classifiers.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Asuncion, A., Newman, D.: UCI Machine Learning Repository (2007), http://www.ics.uci.edu/~mlearn/MLRepository.html
Demšar, J.: Statistical comparisons of classifiers over multiple data sets. Journal of Machine Learning Research 7, 1–30 (2006)
Di Marzio, M., Taylor, C.C.: Boosting kernel density estimates: A bias reduction technique? Biometrika 91(1), 226–233 (2004)
Di Marzio, M., Taylor, C.C.: Kernel density classification and boosting: an l2 analysis. Statistics and Computing 15, 113–123 (2005)
Di Marzio, M., Taylor, C.C.: On boosting kernel density methods for multivariate data: density estimation and classification. Statistical Methods and Applications 14, 163–178 (2005)
Ghosh, A.K., Chaudhuri, P., Sengupta, D.: Classification using kernel density estimates: Multiscale analysis and visualization. Technometrics 48(1), 120–132 (2006)
Kobos, M., Mańdziuk, J.: Multiple-resolution classification with combination of density estimators. Connection Science 23(4), 219–237 (2011)
Kobos, M.: Combination of independent kernel density estimators in classification. In: Ganzha, M., Paprzycki, M. (eds.) Proceedings of the International Multiconference on Computer Science and Information Technology, vol. 4, pp. 57–63 (2009)
Kobos, M., Mańdziuk, J.: Classification Based on Combination of Kernel Density Estimators. In: Alippi, C., Polycarpou, M., Panayiotou, C., Ellinas, G. (eds.) ICANN 2009. LNCS, vol. 5769, pp. 125–134. Springer, Heidelberg (2009)
Kobos, M., Mańdziuk, J.: Classification Based on Multiple-Resolution Data View. In: Diamantaras, K., Duch, W., Iliadis, L.S. (eds.) ICANN 2010. LNCS, vol. 6354, pp. 124–129. Springer, Heidelberg (2010)
Lim, T.S., Loh, W.Y., Shih, Y.S.: A comparison of prediction accuracy, complexity, and training time of thirty-three old and new classification algorithms. Machine Learning 40, 203–228 (2000)
Mangasarian, O., Wolberg, W.: Cancer diagnosis via linear programming. Siam News 23, 1–18 (1990)
Marchette, D.J., Priebe, C.E., Rogers, G.W., Solka, J.L.: Filtered kernel density estimation. Computational Statistics 11(2), 95–112 (1996)
Ripley, B.: Pattern recognition and neural networks datasets collection (1996), http://www.stats.ox.ac.uk/pub/PRNN/
Smyth, P., Wolpert, D.: Linearly combining density estimators via stacking. Machine Learning 36, 59–83 (1999)
Witten, I., Frank, E.: Data Mining: Practical Machine Learning Tools and Techniques, 2nd edn. Elsevier (2005)
Yeh, I.C., Yang, K.J., Ting, T.M.: Knowledge discovery on rfm model using bernoulli sequence. Expert Systems with Applications 36, 5866–5871 (2008)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Kobos, M., Mańdziuk, J. (2012). Bandwidth Selection in Kernel Density Estimators for Multiple-Resolution Classification. In: Rutkowski, L., Korytkowski, M., Scherer, R., Tadeusiewicz, R., Zadeh, L.A., Zurada, J.M. (eds) Artificial Intelligence and Soft Computing. ICAISC 2012. Lecture Notes in Computer Science(), vol 7267. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-29347-4_44
Download citation
DOI: https://doi.org/10.1007/978-3-642-29347-4_44
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-29346-7
Online ISBN: 978-3-642-29347-4
eBook Packages: Computer ScienceComputer Science (R0)