Abstract
The Mixture of Experts (ME) model is one of the most popular ensemble methods used in pattern recognition and machine learning. Despite many studies on the theory and application of the ME model, to our knowledge, its training, testing, and evaluation costs have not been investigated yet. After analyzing the ME model in terms of number of required floating point operations, this paper makes an experimental comparison between the ME model and the recently proposed Mixture of Random Prototype Experts. Experiments have been performed on selected datasets from the UCI machine learning repository. Experimental results confirm the expected behavior of the two ME models, while highlighting that the latter performs better in terms of accuracy and run-time performance.
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Jacobs, R., Jordan, M., Barto, A.: Task decomposition through competition in a modular connectionist architecture: the what and where vision tasks. Tech rep. University of Massachusetts, Amherst, MA (1991)
Jacobs, R., Jordan, M., Nowlan, S., Hinton, G.: Adaptive mixtures of local experts. Neural Computation 3, 79–87 (1991)
Jordan, M.I., Jacobs, R.A.: Hierarchical mixtures of experts and the EM algorithm. Neural Comp. 6, 181–214 (1994)
Murphy, P.M., Aha, D.W.: UCI Repository of Machine Learning Databases, Dept. of Information and Computer Science, Univ. of California, Irvine (1994)
Haykin, S.: Neural Networks: A Comprehensive Foundation, 2nd edn. Prentice-Hall, Englewood Cliffs (1999)
Armano, G., Hatami, N.: Mixture of Random Prototype-Based Local Experts. In: Graña Romay, M., Corchado, E., Garcia Sebastian, M.T. (eds.) HAIS 2010. LNCS (LNAI), vol. 6076, pp. 548–556. Springer, Heidelberg (2010)
Duda, R., Hart, P., Stork, D.: Pattern classification, 2nd edn. John Wiley & Sons, New York (2001)
Hennessy, J., Patterson, D.: Computer architecture: a quantitative approach. Morgan Kaufmann, San Mateo (1990)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Armano, G., Hatami, N. (2011). Run-Time Performance Analysis of the Mixture of Experts Model. In: Burduk, R., Kurzyński, M., Woźniak, M., Żołnierek, A. (eds) Computer Recognition Systems 4. Advances in Intelligent and Soft Computing, vol 95. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-20320-6_18
Download citation
DOI: https://doi.org/10.1007/978-3-642-20320-6_18
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-20319-0
Online ISBN: 978-3-642-20320-6
eBook Packages: EngineeringEngineering (R0)