Skip to main content

Explainability of COVID-19 Classification Models Using Dimensionality Reduction of SHAP Values

  • Conference paper
  • First Online:
Intelligent Systems (BRACIS 2023)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 14195))

Included in the following conference series:

  • 224 Accesses

Abstract

The critical scenario in public health triggered by COVID-19 intensified the demand for predictive models to assist in the diagnosis and prognosis of patients affected by this disease. This work evaluates several machine learning classifiers to predict the risk of COVID-19 mortality based on information available at the time of admission. We also apply a visualization technique based on a state-of-the-art explainability approach which, combined with a dimensionality reduction technique, allows drawing insights into the relationship between the features taken into account by the classifiers in their predictions. Our experiments on two real datasets showed promising results, reaching a sensitivity of up to 84% and an AUROC of 92% (95% CI, [0.89–0.95]).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 59.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 79.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    Ethics committee approval: HCPA 32314720.8.0000.5327, HMV 32314720.8.3001.5330.

  2. 2.

    https://fracpete.github.io/python-weka-wrapper3.

  3. 3.

    https://github.com/dmlc/xgboost.

  4. 4.

    https://github.com/slundberg/shap.

References

  1. Aas, K., Jullum, M., Løland, A.: Explaining individual predictions when features are dependent: accurate approximations to Shapley values. Artif. Intell. 298, 103502 (2021)

    Article  MathSciNet  MATH  Google Scholar 

  2. Alballa, N., Al-Turaiki, I.: Machine learning approaches in Covid-19 diagnosis, mortality, and severity risk prediction: a review. Inf. Med. Unlocked 24, 100564 (2021)

    Article  Google Scholar 

  3. Araújo, D.C., Veloso, A.A., Borges, K.B.G., das Graças Carvalho, M.: Prognosing the risk of Covid-19 death through a machine learning-based routine blood panel: a retrospective study in brazil. IJMEDI 165, 104835 (2022)

    Google Scholar 

  4. Broberg, C.S., Kovacs, A.H., Sadeghi, S., et al.: Covid-19 in adults with congenital heart disease. JACC 77(13), 1644–1655 (2021)

    Article  Google Scholar 

  5. Covino, M., Sandroni, C., Santoro, M., et al.: Predicting intensive care unit admission and death for Covid-19 patients in the emergency department using early warning scores. Resuscitation 156, 84–91 (2020)

    Article  Google Scholar 

  6. Futoma, J., Simons, M., Panch, T., et al.: The myth of generalisability in clinical research and machine learning in health care. Lancet Digit. Health 2(9), 489–492 (2020)

    Article  Google Scholar 

  7. Hall, M.A.: Correlation-based feature selection for machine learning. Ph.D. thesis, The University of Waikato (1999)

    Google Scholar 

  8. Huang, X., Marques-Silva, J.: The inadequacy of Shapley values for explainability. arXiv preprint arXiv:2302.08160 (2023)

  9. Knight, S.R., Ho, A., Pius, R., et al.: Risk stratification of patients admitted to hospital with Covid-19 using the ISARIC WHO clinical characterisation protocol: development and validation of the 4C mortality score. BMJ 370, m3339 (2020)

    Article  Google Scholar 

  10. Lundberg, S.M., Erion, G.G., Lee, S.I.: Consistent individualized feature attribution for tree ensembles. arXiv preprint arXiv:1802.03888 (2018)

  11. Lundberg, S.M., Lee, S.I.: A unified approach to interpreting model predictions. In: Advances in Neural Information Processing Systems, vol. 30 (2017)

    Google Scholar 

  12. Van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. JMLR 9(11), 2579–2605 (2008)

    MATH  Google Scholar 

  13. Miranda de Paiva, B.B., Delfino-Pereira, P., de Andrade, C.M.V., et al.: Effectiveness, explainability and reliability of machine meta-learning methods for predicting mortality in patients with COVID-19: results of the Brazilian COVID-19 registry. medRxiv (2021)

    Google Scholar 

  14. Qin, Z., Zhang, C., Wang, T., Zhang, S.: Cost sensitive classification in data mining. In: Advanced Data Mining and Applications, pp. 1–11 (2010)

    Google Scholar 

  15. Rostami, M., Mansouritorghabeh, H.: D-dimer level in COVID-19 infection: a systematic review. Exp. Rev. Hematol. 13(11), 1265–1275 (2020)

    Article  Google Scholar 

  16. Subudhi, S., Verma, A., Patel, A.B.: Prognostic machine learning models for Covid-19 to facilitate decision making. IJCP 74(12), e13685 (2020)

    Google Scholar 

  17. Wynants, L., Van Calster, B., Collins, G.S., et al.: Prediction models for diagnosis and prognosis of COVID-19: systematic review and critical appraisal. bmj 369, m1328 (2020)

    Article  Google Scholar 

  18. Yadaw, A.S., Li, Y., Bose, S., et al.: Clinical features of COVID-19 mortality: development and validation of a clinical prediction model. Lancet Digt. Health 2(10), E516–E525 (2020)

    Article  Google Scholar 

  19. Yang, J.: Fast TreeSHAP: accelerating SHAP value computation for trees. arXiv preprint arXiv:2109.09847 (2021)

  20. Zhao, Z., Chen, A., Hou, W., et al.: Prediction model and risk scores of ICU admission and mortality in COVID-19. PLoS ONE 15(7), e0236618 (2020)

    Article  Google Scholar 

Download references

Acknowledgment

This work has been financed in part by CAPES Finance Code 001 and CNPq/Brazil.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Daniel Matheus Kuhn .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Kuhn, D.M., de Loreto, M.S., Recamonde-Mendoza, M., Comba, J.L.D., Moreira, V.P. (2023). Explainability of COVID-19 Classification Models Using Dimensionality Reduction of SHAP Values. In: Naldi, M.C., Bianchi, R.A.C. (eds) Intelligent Systems. BRACIS 2023. Lecture Notes in Computer Science(), vol 14195. Springer, Cham. https://doi.org/10.1007/978-3-031-45368-7_27

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-45368-7_27

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-45367-0

  • Online ISBN: 978-3-031-45368-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics