Abstract
The Fisher kernel has good statistical properties. However, from a practical point of view, the necessary distributional assumptions complicate the applicability. We approach the solution to this problem with the NMF (Non-negative Matrix Factorization) methods, which with adequate normalization conditions, provide stochastic matrices. Using the Bregman divergence as the objective function, formally equivalent solutions appear for the specific forms of the functionals involved. We show that simply by taking these results and plug-in into the general expression of the NMF kernel, obtained with purely algebraic techniques, without any assumptions about the distribution of the parameters, the properties of the Fisher kernel hold, and it is a convenient procedure to use this kernel the situations in which they are needed we derive the expression of the information matrix of Fisher. In this work, we have limited the study to the Gaussian metrics, KL (Kullback-Leibler), and I-divergence.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Aggarwal, C.C., Clustering, C.R.D.: Algorithms and Applications. CRC Press Taylor and Francis Group, Boca Raton (2014)
Amari, S.I.: Information geometry of the EM and EM algorithms for neural networks. Neural Netw. 8(9), 1379–1408 (1995)
Aronszajn, N.: Theory of reproducing kernels. Trans. Am. Math. Soc. 68(3), 337–404 (1950)
Banerjee, A., Merugu, S., Dhillon, I.S., Ghosh, J., Lafferty, J.: Clustering with Bregman divergences. J. Mach. Learn. Res. 6(10), 1705–174 (2005)
Chappelier, J.-C., Eckard, E.: PLSI: the true fisher kernel and beyond. In: Buntine, W., Grobelnik, M., Mladenić, D., Shawe-Taylor, J. (eds.) ECML PKDD 2009. LNCS (LNAI), vol. 5781, pp. 195–210. Springer, Heidelberg (2009). https://doi.org/10.1007/978-3-642-04180-8_30
Chen, J.C.: The nonnegative rank factorizations of nonnegative matrices. Linear Algebra Appl. 62, 207–217 (1984)
Ding, C., Li, T., Peng, W.: On the equivalence between non-negative matrix factorization and probabilistic latent semantic indexing. Comput. Stat. Data Anal. 52(8), 3913–3927 (2008)
Dua, D., Graff, C.: UCI machine learning repository. University of California, Irvine, School of Information and Computer Sciences (2017). http://archive.ics.uci.edu/ml
Figuera, P., García Bringas, P.: On the probabilistic latent semantic analysis generalization as the singular value decomposition probabilistic image. J. Stat. Theory Appl. 19, 286–296 (2020)
Figuera, P., Bringas, P.G.: A non-parametric fisher kernel. In: Sanjurjo González, H., Pastor López, I., García Bringas, P., Quintián, H., Corchado, E. (eds.) HAIS 2021. LNCS (LNAI), vol. 12886, pp. 448–459. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-86271-8_38
Gudovskiy, D., Hodgkinson, A., Yamaguchi, T., Tsukizawa, S.: Deep active learning for biased datasets via fisher kernel self-supervision. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9041–9049 (2020)
Hofmann, T.: Learning the similarity of documents: an information-geometric approach to document retrieval and categorization. In: Advances in Neural Information Processing Systems, pp. 914–920 (2000)
Hofmann, T., Schölkopf, B., Smola, A.J.: Kernel methods in machine learning. Ann. Stat. 36, 1171–1220 (2008)
Jaakkola, T., Diekhans, M., Haussler, D.: A discriminative framework for detecting remote protein homologies. J. Comput. Biol. 7(1–2), 95–114 (2000)
Jaakkola, T.S., Haussler, D., et al.: Exploiting generative models in discriminative classifiers. In: Advances in Neural Information Processing Systems, pp. 487–493 (1999)
Kim, D.: KFDA: kernel fisher discriminant analysis (2017). https://CRAN.R-project.org/package=kfda, R package version 1.0.0
Langrené, N., Warin, X.: Fast and stable multivariate kernel density estimation by fast sum updating. J. Comput. Graph. Stat. 28(3), 596–608 (2019)
Lee, H., Cichocki, A., Choi, S.: Kernel nonnegative matrix factorization for spectral EEG feature extraction. Neurocomputing 72(13–15), 3182–3190 (2009)
Mercer, J.: XVI. Functions of positive and negative type, and their connection the theory of integral equations. Philos. Trans. R. Soc. Lond. Ser. A Containing Papers Math. Phys. Character 209(441–458), 415–446 (1909)
Rao, C.R.: Differential metrics in probability spaces. Diff. Geom. Stat. Inference 10, 217–240 (1987)
Salazar, D., Rios, J., Aceros, S., Flórez-Vargas, O., Valencia, C.: Kernel joint non-negative matrix factorization for genomic data. IEEE Access 9, 101863–101875 (2021)
Scholkopft, B., Mullert, K.R.: Fisher discriminant analysis with kernels. Neural Netw. Signal Process. IX 1(1), 1 (1999)
Sharma, A., Paliwal, K.K.: Rotational linear discriminant analysis technique for dimensionality reduction. IEEE Trans. Knowl. Data Eng. 20(10), 1336–1347 (2008)
Tsuda, K., Akaho, S., Kawanabe, M., Müller, K.R.: Asymptotic properties of the fisher kernel. Neural Comput. 16(1), 115–137 (2004)
Vapnik, V.N.: An overview of statistical learning theory. IEEE Trans. Neural Netw. 10(5), 988–999 (1999)
Zhang, D., Zhou, Z.-H., Chen, S.: Non-negative matrix factorization on kernels. In: Yang, Q., Webb, G. (eds.) PRICAI 2006. LNCS (LNAI), vol. 4099, pp. 404–412. Springer, Heidelberg (2006). https://doi.org/10.1007/978-3-540-36668-3_44
Zhou, Y., Shi, J., Zhu, J.: Nonparametric score estimators. In: International Conference on Machine Learning, pp. 11513–11522. PMLR (2020)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Figuera, P., Cuzzocrea, A., Bringas, P.G. (2022). Generalized Fisher Kernel with Bregman Divergence. In: García Bringas, P., et al. Hybrid Artificial Intelligent Systems. HAIS 2022. Lecture Notes in Computer Science(), vol 13469. Springer, Cham. https://doi.org/10.1007/978-3-031-15471-3_17
Download citation
DOI: https://doi.org/10.1007/978-3-031-15471-3_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-15470-6
Online ISBN: 978-3-031-15471-3
eBook Packages: Computer ScienceComputer Science (R0)