Skip to main content
Log in

Parallel Approach for Ensemble Learning with Locally Coupled Neural Networks

  • Published:
Neural Processing Letters Aims and scope Submit manuscript

Abstract

Ensemble learning has gained considerable attention in different tasks including regression, classification and clustering. Adaboost and Bagging are two popular approaches used to train these models. The former provides accurate estimations in regression settings but is computationally expensive because of its inherently sequential structure, while the latter is less accurate but highly efficient. One of the drawbacks of the ensemble algorithms is the high computational cost of the training stage. To address this issue, we propose a parallel implementation of the Resampling Local Negative Correlation (RLNC) algorithm for training a neural network ensemble in order to acquire a competitive accuracy like that of Adaboost and an efficiency comparable to that of Bagging. We test our approach on both synthetic and real datasets from the UCI and Statlib repositories for the regression task. In particular, our fine-grained parallel approach allows us to achieve a satisfactory balance between accuracy and parallel efficiency.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Blake C, Newman D, Merz C (1998) UCI repository of machine learning databases

  2. Breiman L (1996) Bagging predictors. Mach Learn 24(2): 123–140

    MATH  MathSciNet  Google Scholar 

  3. Breshears C (2009) The art of concurrency: a thread monkey’s guide to writing parallel applications. O’Reilly Media

  4. Brown G (2003) Diversity in neural network ensembles. PhD Thesis, School of Computer Science, University of Birmingham

  5. Chu C, Kim S, Lin Y, Yu Y, Bradski GR, Ng A, Olukotun K (2006) Map-reduce for machine learning on multicore. In: NIPS MIT Press, pp 281–288

  6. Dietterich T, Fisher D (2000) An experimental comparison of three methods for constructing ensembles of decision trees. In: Bagging, boosting, and randomization. Machine learning. pp 139–157

  7. Drucker H (1997) Improving regressors using boosting techniques. In: Fourteenth international conference on machine learning. pp 107–115

  8. Estévez P, Paugam-Moisy H, Puzenat D, Ugarte M (2002) A scalable parallel algorithm for training a hierarchical mixture of neural experts. Parallel Comput, 28(6): 861–891

    Article  MathSciNet  Google Scholar 

  9. Friedman J (1991) Multivariate adaptive regression splines. Ann Stat 19(1): 1–67

    Article  MATH  Google Scholar 

  10. Galtier V, Pietquin O, Vialle S (2007) Adaboost parallelization on pc clusters with virtual shared memory for fast feature selection. In: Proceedings of the IEEE international conference on signal processing and communication. November 2007. Dubai, United Arab Emirates, pp 165–168

  11. Grama A, Karypis G, Kumar V, Gupta A (2003) Introduction to parallel computing. 2nd edn. Addison Wesley, Boston, MA

    Google Scholar 

  12. Grandvalet Y (2004) Bagging equalizes influence. Mach Learn 55(3): 251–270

    Article  MATH  Google Scholar 

  13. Gropp W (2002) Mpich2: a new start for mpi implementations. In: Proceedings of the 9th European PVM/MPI Users’ Group Meeting on recent advances in parallel virtual machine and message passing interface. Springer-Verlag, London, UK, p 7

  14. Hastie T, Tibshirani R, Friedman J (2003) The elements of statistical learning, Chap. 10. Springer, New York

    Google Scholar 

  15. Kargupta, H, Joshi, A, Sivakumar, K, Yesha, Y (eds) (2004) Data mining: next generation challenges and future directions. AAAI Press, Menlo Park, CA

    Google Scholar 

  16. Lazarevic A, Obradovic Z (2002) Boosting algorithms for parallel and distributed learning. Distrib Parallel Databases 11(2): 203–229

    Article  MATH  Google Scholar 

  17. Martínez-Muñoz G, Suárez A (2010) Out-of-bag estimation of the optimal sample size in bagging. Pattern Recogn 43(1): 143–152

    Article  MATH  Google Scholar 

  18. Merler S, Caprile B, Furlanello C (2007) Parallelizing adaboost by weights dynamics. Comput Stat Data Anal, 51: 2487–2498

    Article  MATH  MathSciNet  Google Scholar 

  19. Ñanculef R, Valle C, Allende H, Moraga C (2006) Ensemble learning with local diversity. In: ICANN (1), Lecture Notes in Computer Science, vol 4131. Springer, pp 570–577

  20. Ñanculef R, Valle C, Allende H, Moraga C (2006) Local negative correlation with resampling. In: IDEAL, Lecture Notes in Computer Science, vol 4224. Springer, pp 570–577

  21. Schapire R (1999) A short introduction to boosting. In Proceedings of the sixteenth international joint conference on artificial intelligence. pp 771–780

  22. Shrestha D, Solomatine D (2006) Experiments with adaboost.rt, an improved boosting scheme for regression. Neural Comput 18(7): 1678–1710

    Article  MATH  Google Scholar 

  23. Vin T, Seng M, Kuan N, Haron F (2005) A framework for grid-based neural networks. In: Distributed frameworks for multimedia applications, International conference on, pp 246–253

  24. Vlachos P (2000) Statlib project repository. Department of Statistics, Carnegie Mellon University

  25. Woodsend K, Gondzio J (2009) Hybrid MPI/OpenMP parallel linear support vector machine training. J Mach Learn Res 10: 1937–1953

    MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Carlos Valle.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Valle, C., Saravia, F., Allende, H. et al. Parallel Approach for Ensemble Learning with Locally Coupled Neural Networks. Neural Process Lett 32, 277–291 (2010). https://doi.org/10.1007/s11063-010-9157-6

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11063-010-9157-6

Keywords

Navigation