Skip to main content

Empirical Comparison Between Cross-Validation and Mutation-Validation in Model Selection

  • Conference paper
  • First Online:
Advances in Intelligent Data Analysis XXII (IDA 2024)

Abstract

Mutation validation (MV) is a recently proposed approach for model selection, garnering significant interest due to its unique characteristics and potential benefits compared to the widely used cross-validation (CV) method. In this study, we empirically compared MV and k-fold CV using benchmark and real-world datasets. By employing Bayesian tests, we compared generalization estimates yielding three posterior probabilities: practical equivalence, CV superiority, and MV superiority. We also evaluated the differences in the capacity of the selected models and computational efficiency. We found that both MV and CV select models with practically equivalent generalization performance across various machine learning algorithms and the majority of benchmark datasets. MV exhibited advantages in terms of selecting simpler models and lower computational costs. However, in some cases MV selected overly simplistic models leading to underfitting and showed instability in hyperparameter selection. These limitations of MV became more evident in the evaluation of a real-world neuroscientific task of predicting sex at birth using brain functional connectivity.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    https://scikit-learn.org/stable/modules/generated/sklearn.model_selection.KFold.html.

  2. 2.

    https://scikit-learn.org/stable/modules/feature_selection.html.

References

  1. xcpengine-container 1.0.1. https://pypi.org/project/xcpengine-container/

  2. Barbiero, P., Squillero, G., Tonda, A.P.: Modeling generalization in machine learning: a methodological and computational study. CoRR abs/2006.15680 (2020)

    Google Scholar 

  3. Corani, G., Benavoli, A.: A Bayesian approach for comparing cross-validated algorithms on multiple data sets. Mach. Learn. 100, 285–304 (2015)

    Article  MathSciNet  Google Scholar 

  4. Corani, G., Benavoli, A., Demšar, J., Mangili, F., Zaffalon, M.: Statistical comparison of classifiers through Bayesian hierarchical modelling. Mach. Learn. 106(11), 1817–1837 (2017)

    Article  MathSciNet  Google Scholar 

  5. Dua, D., Graff, C.: UCI machine learning repository (2017). http://archive.ics.uci.edu/ml

  6. Esteban, O., et al.: fMRIPrep: a robust preprocessing pipeline for functional MRI (2022)

    Google Scholar 

  7. Feldman, V., Frostig, R., Hardt, M.: The advantages of multiple classes for reducing overfitting from test set reuse. CoRR abs/1905.10360 (2019)

    Google Scholar 

  8. Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003)

    Google Scholar 

  9. Hastie, T., Tibshirani, R., Friedman, J.: The Elements of Statistical Learning: Data Mining, Inference, and Prediction. Springer Series in Statistics, Springer, New York (2009). https://doi.org/10.1007/978-0-387-21606-5

    Book  Google Scholar 

  10. Kohavi, R.: A Study of Cross-validation and Bootstrap for Accuracy Estimation and Model Selection, pp. 1137–1143. IJCAI 1995, Morgan Kaufmann Publishers Inc., San Francisco, CA, USA (1995)

    Google Scholar 

  11. Kruschke, J.: Bayesian estimation supersedes the t test. J. Exp. Psychol. General 142, 573–603 (2012)

    Article  Google Scholar 

  12. Mitchell, T.M.: Machine Learning. McGraw-hill, New York (1997)

    Google Scholar 

  13. Raschka, S.: Model evaluation, model selection, and algorithm selection in machine learning. CoRR abs/1811.12808 (2018)

    Google Scholar 

  14. Schaefer, A., et al.: Local-Global Parcellation of the Human Cerebral Cortex from Intrinsic Functional Connectivity MRI. Cerebral Cortex (2018)

    Google Scholar 

  15. Snoek, L., et al.: The Amsterdam open MRI collection, a set of multimodal MRI datasets for individual difference analyses. Sci. Data 8(1), 85 (2021)

    Article  Google Scholar 

  16. Vanschoren, J., van Rijn, J.N., Bischl, B., Torgo, L.: OpenML: networked science in machine learning. SIGKDD Explor. 15(2), 49–60 (2013)

    Article  Google Scholar 

  17. Weis, S., Patil, K.R., Hoffstaedter, F., Nostro, A., Yeo, B.T.T., Eickhoff, S.B.: Sex classification by resting state brain connectivity. Cereb. Cortex 30(2), 824–835 (2019)

    Article  Google Scholar 

  18. Zhang, J.M., Harman, M., Guedj, B., Barr, E.T., Shawe-Taylor, J.: Model validation using mutated training labels: an exploratory study. Neurocomput. 539(C), 126116 (2023). https://doi.org/10.1016/j.neucom.2023.02.042

Download references

Acknowledgements

This work was partly supported by the Helmholtz Portfolio Theme “Supercomputing and Modelling for the Human Brain” and by the Max Planck School of Cognition supported by the Federal Ministry of Education and Research (BMBF) and the Max Planck Society (MPG).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kaustubh R. Patil .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yu, J., Hamdan, S., Sasse, L., Morrison, A., Patil, K.R. (2024). Empirical Comparison Between Cross-Validation and Mutation-Validation in Model Selection. In: Miliou, I., Piatkowski, N., Papapetrou, P. (eds) Advances in Intelligent Data Analysis XXII. IDA 2024. Lecture Notes in Computer Science, vol 14642. Springer, Cham. https://doi.org/10.1007/978-3-031-58553-1_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-58553-1_5

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-58555-5

  • Online ISBN: 978-3-031-58553-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics