Abstract
Due to the fast learning speed, simplicity of implementation and minimal human intervention, extreme learning machine has received considerable attentions recently, mostly from the machine learning community. Generally, extreme learning machine and its various variants focus on classification and regression problems. Its potential application in analyzing censored time-to-event data is yet to be verified. In this study, we present an extreme learning machine ensemble to model right-censored survival data by combining the Buckley-James transformation and the random forest framework. According to experimental and statistical analysis results, we show that the proposed model outperforms popular survival models such as random survival forest, Cox proportional hazard models on well-known low-dimensional and high-dimensional benchmark datasets in terms of both prediction accuracy and time efficiency.
Similar content being viewed by others
References
David C R (1972) Regression models and life tables (with discussion). J R Stat Soc 34:187–220
Cox D R, Oakes D (1984) Analysis of survival data, vol 21. CRC Press, Cleveland
Tibshirani R et al (1997) The lasso method for variable selection in the cox model. Stat Med 16(4):385–395
Gui J, Li H (2005) Penalized cox regression analysis in the high-dimensional and low-sample size settings, with applications to microarray gene expression data. Bioinformatics 21(13):3001–3008
Simon N, Friedman J H, Hastie T, Tibshirani R (2011) Regularization paths for cox’s proportional hazards model via coordinate descent. J Stat Softw 39(5):1–13
Faraggi D, Simon R (1995) A neural network model for survival data. Stat Med 14(1):73–82
Biganzoli E, Boracchi P, Mariani L, Marubini E (1998) Feed forward neural networks for the analysis of censored survival data: a partial logistic regression approach. Stat Med 17(10):1169– 1186
LeBlanc M, Crowley J (1995) A review of tree-based prognostic models. In: Recent advances in clinical trial design and analysis. Springer, Berlin, pp 113–124
Bou-Hamad I, Larocque D, Ben-Ameur H (2011) A review of survival trees. Stat Surv 5:44–71
Hothorn T, Bühlmann P, Dudoit S, Molinaro A, Van Der Laan M J (2006) Survival ensembles. Biostatistics 7(3):355–373
Ishwaran H, Kogalur U B, Blackstone E H, Lauer M S (2008) Random survival forests. Ann Appl Stat 2(3):841–860
Lu W, Li L (2008) Boosting method for nonlinear transformation models with censored survival data. Biostatistics 9(4):658–667
Ravdin P M, Clark G M (1992) A practical application of neural network analysis for predicting outcome of individual breast cancer patients. Breast Cancer Res Treat 22(3):285–293
Ebell M H (1993) Artificial neural networks for predicting failure to survive following in-hospital cardiopulmonary resuscitation. J Fam Pract 36(3):297–304
Liestbl K, Andersen P K, Andersen U (1994) Survival analysis and neural nets. Stat Med 13(12):1189–1200
Lisboa P J, Wong H, Harris P, Swindell R (2003) A bayesian neural network approach for modelling censored data with an application to prognosis after surgery for breast cancer. Artif Intell Med 28(1):1–25
Biganzoli E M, Boracchi P, Ambrogi F, Marubini E (2006) Artificial neural network for the joint modelling of discrete cause-specific hazards. Artif Intell Med 37(2):119–130
Ambrogi F, Lama N, Boracchi P, Biganzoli E (2007) Selection of artificial neural network models for survival analysis with genetic algorithms. Comput Stat Data Anal 52(1):30–42
Lisboa P J, Etchells T A, Jarman I H, Aung M H, Chabaud S, Bachelot T, Perol D, Gargi T, Bourdès V, Bonnevay S et al (2008) Time-to-event analysis with artificial neural networks: an integrated analytical and rule-based study for breast cancer. Neural Netw 21(2):414–426
Xiang A, Lapuerta P, Ryutov A, Buckley J, Azen S (2000) Comparison of the performance of neural network methods and cox regression for censored survival data. Computat Stat Data Anal 34(2):243–257
Huang G -B, Zhu Q -Y, Siew C -K (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1):489– 501
Huang G -B, Zhou H, Ding X, Zhang R (2012) Extreme learning machine for regression and multiclass classification. IEEE Trans Syst Man Cybern Part B (Cybern) 42(2):513–529
Huang G -B, Chen L, Siew C K et al (2006) Universal approximation using incremental constructive feedforward networks with random hidden nodes. IEEE Trans Neural Netw 17(4):879–892
Wang Y, Cao F, Yuan Y (2011) A study on effectiveness of extreme learning machine. Neurocomputing 74(16):2483–2490
Buckley J, James I (1979) Linear regression with censored data. Biometrika 66:429–436
Wang S, Nan B, Zhu J, Beer D G (2008) Doubly penalized buckley–james method for survival data with high-dimensional covariates. Biometrics 64(1):132–140
Breiman L (2001) Random forests. Mach Learn 45(1):5–32
Kaplan E L, Meier P (1958) Nonparametric estimation from incomplete observations. J Amer Stat Assoc 53(282):457– 481
Breiman L (1996) Bagging predictors. Mach Learn 24(2):123–140
Ho T K (1998) The random subspace method for constructing decision forests. IEEE Trans Pattern Anal Mach Intell 20(8):832– 844
Breiman L, Friedman J, Stone C J, Olshen R A (1984) Classification and regression trees. CRC Press, Cleveland
Fernández-Delgado M, Cernadas E, Barro S, Amorim D (2014) Do we need hundreds of classifiers to solve real world classification problems. J Mach Learn Res 15(1):3133–3181
Ren Y, Zhang L, Suganthan P N (2016) Ensemble classification and regression-recent developments, applications and future directions [review article]. IEEE Comput Intell Mag 11(1):41– 53
Jia L, Liao S (2009) Accurate probabilistic error bound for eigenvalues of kernel matrix. In: Asian conference on machine learning. Springer, Berlin, pp 162–175
Polikar R (2006) Ensemble based systems in decision making. Circs Syst Mag IEEE 6(3):21–45
Melville P, Mooney R J (2003) Constructing diverse classifier ensembles using artificial training examples. In: Proceedings of the 18th international joint conference on artificial intelligence, IJCAI’03. Morgan Kaufmann Publishers Inc., San Francisco, pp 505–510
Harrell F E, Lee K L, Mark D B (1996) Tutorial in biostatistics multivariable prognostic models: issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors. Stat Med 15:361–387
Dietterich T G (1998) Approximate statistical tests for comparing supervised classification learning algorithms. Neural Comput 10(7):1895–1923
Demṡar J (2006) Statistical comparisons of classifiers over multiple data sets. J Mach Learn Res 7:1–30
Hsu C W, Chang CC, Lin CJ (2016) A practical guide to support vector classification, Tech. rep., National Taiwan University. http://www.csie.ntu.edu.tw/cjlin/papers/guide/guide.pdf. Accessed 22 May 2017
R Core Team (2016) R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna. https://www.R-project.org/
Pohlert T (2014) The pairwise multiple comparison of mean ranks package (PMCMR). R package. http://CRAN.R-project.org/package=PMCMR. Accessed 22 May 2017
Wright M, Ziegler A (2017) ranger: a fast implementation of random forests for high dimensional data in c++ and r. J Stat Softw 77(1):1–17. https://doi.org/10.18637/jss.v077.i01
Ridgeway G (2017) gbm: generalized boosted regression models. R package. http://CRAN.R-project.org/package=gbm Accessed 22 May 2017
Mersmann O (2015) microbenchmark: accurate timing functions. R package. https://CRAN.R-project.org/package=microbenchmark. Accessed 22 May 2017
Deng W, Zheng Q, Chen L (2009) Regularized extreme learning machine. In: IEEE Symposium on computational intelligence and data mining, 2009. CIDM’09. IEEE, New York, pp 389– 395
Liu N, Wang H (2010) Ensemble based extreme learning machine. IEEE Signal Process Lett 17(8):754–757
Lu H-J, An C-L, Zheng E-H, Lu Y (2014) Dissimilarity based ensemble of extreme learning machine for gene expression data classification. Neurocomputing 128:22–30
Yu Q, Van Heeswijk M, Miche Y, Nian R, He B, Séverin E, Lendasse A (2014) Ensemble delta test-extreme learning machine (dt-elm) for regression. Neurocomputing 129:153– 158
Acknowledgements
This work was supported in part by National Social Science Foundation of China (17BTJ019), Social Science Foundation for Young Scholars of Ministry of Education of China (15YJCZH166), Hunan Provincial Social Science Foundation of China (16YBA367), the scholarship from China Scholarship Council (CSC201606375129), China Postdoctoral Science Foundation (2017M612574) and Postgraduates Education Reform Fund (2016JGB25) at Central South University, China.
The funders had no role in the preparation of this article.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Wang, H., Wang, J. & Zhou, L. A survival ensemble of extreme learning machine. Appl Intell 48, 1846–1858 (2018). https://doi.org/10.1007/s10489-017-1063-4
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-017-1063-4