Abstract
We generalize traditional goals of clustering towards distinguishing components in a non-parametric mixture model. The clusters are not necessarily based on point locations, but on higher order criteria. This framework can be implemented by embedding probability distributions in a Hilbert space. The corresponding clustering objective is very general and relates to a range of common clustering concepts.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Aghagolzadeh, M., Soltanian-Zadeh, H., Araabi, B., Aghagolzadeh, A.: A hierarchical clustering based on maximizing mutual information. In: ICIP (2007)
Arora, S., Kannan, R.: Learning mixtures of arbitrary Gaussians. In: STOC (2001)
Chaudhuri, K., Rao, S.: Beyond Gaussians: Spectral methods for learning mixtures of heavy-tailed distributions. In: COLT (2008)
Erdogmus, D., Principe, J.C.: From linear adaptive filtering to nonlinear information processing. IEEE Signal Processing Magazine 23(6), 14–33 (2006)
Fuglede, B., Topsøe, F.: Jensen-Shannon divergence and Hilbert space embedding. In: Proc. of the Int. Symp. on Information Theory, ISIT (2004)
Gokcay, E., Principe, J.C.: Information theoretic clustering. IEEE Transactions on Pattern Analysis and Machine Intelligence 24(2), 158–170 (2002)
Gretton, A., Bousquet, O., Smola, A., Schölkopf, B.: Measuring statistical dependence with Hilbert-Schmidt norms. In: Jain, S., Simon, H.U., Tomita, E. (eds.) ALT 2005. LNCS (LNAI), vol. 3734, pp. 63–77. Springer, Heidelberg (2005)
Gretton, A., Borgwardt, K.M., Rasch, M., Schölkopf, B., Smola, A.: A kernel method for the two-sample problem. In: NIPS (2006)
Gretton, A., Borgwardt, K., Rasch, M., Schölkopf, B., Smola, A.: A kernel method for the two-sample problem. Tech. Report 157, MPI for Biol. Cyb. (2008)
Jenssen, R., Erdogmus, D., Principe, J., Eltoft, T.: The laplacian pdf distance: a cost function for clustering in a kernel feature space. In: NIPS (2004)
Lin, J.: Divergence measures based on the Shannon entropy. IEEE Transaction on Information Theory 37(1) (1991)
Newman, M.E.J., Girvan, M.: Finding and evaluating community structure in networks. Physical review E 69 (2004)
Renyi, A.: On measures of entropy and information. In: Proc. Fourth Berkeley Symp. Math., Statistics and Probability, pp. 547–561 (1960)
Rose, K.: A mapping approach to rate-distortion computation and analysis. IEEE Transactions on Information Theory 40(6), 1939–1952 (1994)
Song, L., Smola, A., Gretton, A., Borgwardt, K.: A dependence maximization view of clustering. In: 24th International Conference on Machine Learning, ICML (2007)
Sriperumbudur, B.K., Gretton, A., Fukumizu, K., Lanckriet, G., Schölkopf, B.: Injective Hilbert space embeddings of probability measures. In: COLT (2008)
Wächter, A., Biegler, L.T.: On the implementation of a primal-dual interior point filter line search algorithm for large-scale nonlinear programming. Mathematical Programming 106(1), 25–57 (2006)
Xu, L., Neufeld, J., Larson, B., Schuurmans, D.: Maximum margin clustering. In: NIPS (2005)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2009 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Jegelka, S., Gretton, A., Schölkopf, B., Sriperumbudur, B.K., von Luxburg, U. (2009). Generalized Clustering via Kernel Embeddings. In: Mertsching, B., Hund, M., Aziz, Z. (eds) KI 2009: Advances in Artificial Intelligence. KI 2009. Lecture Notes in Computer Science(), vol 5803. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-04617-9_19
Download citation
DOI: https://doi.org/10.1007/978-3-642-04617-9_19
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-04616-2
Online ISBN: 978-3-642-04617-9
eBook Packages: Computer ScienceComputer Science (R0)