Abstract
Hidden Markov models (HMMs) are widely used for modeling multivariate time series data. However, all collected data is not always useful for distinguishing between states. In these situations, feature selection should be implemented to save the expense of collecting and processing low utility data. Feature selection for HMMs has been studied but most existing methods assume that the observed data follows a Gaussian distribution. In this paper, a method for simultaneously estimating parameters and selecting features for an HMM with discrete observations is presented. The presented method is an extension of the feature saliency HMM which was originally developed to incorporate feature cost into the feature selection process. Expectation-maximization algorithms are derived for features following a Poisson distribution and features following a discrete non-parametric distribution. The algorithms are evaluated on synthetic data sets and a real-world event detection data set that is composed of both Poisson and non-parametric features.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Bian, J., Zhou, X.: Hidden Markov models in bioinformatics: SNV inference from next generation sequence. In: Hidden Markov Models, pp. 123–133. Springer (2017)
Candanedo, L.M., Feldheim, V., Deramaix, D.: A methodology based on hidden Markov models for occupancy detection and a case study in a low energy residential building. Energy Build. 148, 327–341 (2017)
Karg, M., Kulić, D.: Modeling movement primitives with hidden Markov models for robotic and biomedical applications. In: Hidden Markov Models, pp. 199–213. Springer (2017)
Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003)
Li, J., Cheng, K., Wang, S., Morstatter, F., Trevino, R.P., Tang, J., Liu, H.: Feature selection: a data perspective. ACM Comput. Surv. (CSUR) 50(6), 94 (2017)
Adams, S., Beling, P.A.: A survey of feature selection methods for Gaussian mixture models and hidden Markov models. In: Artificial Intelligence Review pp. 1–41 (2017)
Adams, S., Beling, P.A., Cogill, R.: Feature selection for hidden Markov models and hidden semi-Markov models. IEEE Access 4, 1642–1657 (2016)
Law, M.H., Figueiredo, M.A., Jain, A.K.: Simultaneous feature selection and clustering using mixture models. IEEE Trans. Pattern Anal. Mach. Intell. 26(9), 1154–1166 (2004)
Zhu, H., He, Z., Leung, H.: Simultaneous feature and model selection for continuous hidden Markov models. IEEE Signal Process. Lett. 19(5), 279–282 (2012)
Rabiner, L.R.: A tutorial on hidden Markov models and selected applications in speech recognition. Proc. IEEE 77(2), 257–286 (1989)
Gauvain, J.L., Lee, C.H.: Maximum a posteriori estimation for multivariate gaussian mixture observations of Markov chains. IEEE Trans. Speech and Audio Process. 2(2), 291–298 (1994)
Lichman, M.: UCI machine learning repository (2013). http://archive.ics.uci.edu/ml
Ihler, A., Hutchins, J., Smyth, P.: Adaptive event detection with time-varying Poisson processes. In: Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 207–216. ACM (2006)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Adams, S., Beling, P.A. (2020). Feature Selection for Hidden Markov Models with Discrete Features. In: Bi, Y., Bhatia, R., Kapoor, S. (eds) Intelligent Systems and Applications. IntelliSys 2019. Advances in Intelligent Systems and Computing, vol 1037. Springer, Cham. https://doi.org/10.1007/978-3-030-29516-5_7
Download citation
DOI: https://doi.org/10.1007/978-3-030-29516-5_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-29515-8
Online ISBN: 978-3-030-29516-5
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)