Abstract
In recent years, there has been an explosion in the amount of patient Electronic Health Records (EHR) made publicly available. This presents an opportunity to create predictive models that leverage the large amount of data to help guide healthcare worker’s decision-making capacity. However, Patient EHR data is often high-dimensional, sparse, temporal and multimodal. This presents an issue for predictive modeling with Machine Learning models, as many ML model archetypes are not built to handle these types of data. The temporality of EHR data also presents a complicating factor, as it contains multiple time series at different resolutions, which most ML models are not built to handle. This brings us to the topic of patient representation, which is the process of converting this raw EHR data into a dense, mathematical representation. Previous work in this field, however, has not leveraged the full potential of the data, since they opt to only deal with a single modality of data, or do not leverage the temporality of the data. In this paper, we attempt to create a network that creates a multimodal representation of EHR data by modeling it as a multiple sparse time series fusion task. We leverage transformers for sparse time series data, using a custom time-based positional encoding. We then fuse the data into a low-dimensional vector, which serves as our representation. We train the model on 2 separate tasks – Mortality prediction using a classification head, and we attempt to leverage a form of adversarial learning to improve the quality of the representation. We show that the patient representation extracted is meaningful and useful for downstream classification tasks. To our knowledge, this is the first attempt to leverage both adversarial learning and multimodality to create a patient representation. Source code can be found at: https://github.com/BharathSShankar/Patient-Rep-UROPS.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Holroyd-Leduc, J.M., Lorenzetti, D., Straus, S.E., Sykes, L., Quan, H.: The impact of the electronic medical record on structure, process, and outcomes within primary care: a systematic review of the evidence. J. Am. Med. Inform. Assoc. 18(6), 732–737 (2011). https://doi.org/10.1136/amiajnl-2010-000019
Jiang, F., et al.: Artificial intelligence in healthcare: past, present and future. Stroke Vasc. Neurol. 2(4), 230–243 (2017). https://doi.org/10.1136/svn-2017-000101
Norgeot, B., Glicksberg, B.S., Butte, A.J.: A call for deep-learning healthcare. Nat. Med. 25(1), 14–15 (2019). https://doi.org/10.1038/s41591-018-0320-3
Alloghani, M., Baker, T., Al-Jumeily, D., Hussain, A., Mustafina, J., Aljaaf, A.J.: Prospects of machine and deep learning in analysis of vital signs for the improvement of healthcare services. In: Yang, X.-S., He, X.-S. (eds.) Nature-Inspired Computation in Data Mining and Machine Learning, pp. 113–136. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-28553-1_6
Che, C., Zhang, P., Zhu, M., Qu, Y., Jin, B.: Constrained transformer network for ECG signal processing and arrhythmia classification. BMC Med. Inform. Decis. Mak. 21(1), 184 (2021). https://doi.org/10.1186/s12911-021-01546-2
Hosseini, M.-P., Tran, T.X., Pompili, D., Elisevich, K., Soltanian-Zadeh, H.: Deep learning with edge computing for localization of epileptogenicity using multimodal rs-fMRI and EEG big data. In: 2017 IEEE International Conference on Autonomic Computing (ICAC), pp. 83–92 (2017).https://doi.org/10.1109/ICAC.2017.41
Lai, M.: Deep Learning for Medical Image Segmentation (2015). http://arxiv.org/abs/1505.02000
Si, Y., et al.: Deep representation learning of patient data from Electronic Health Records (EHR): a systematic review. J. Biomed. Inform. 115, 103671 (2021). https://doi.org/10.1016/j.jbi.2020.103671
Che, Z., Kale, D., Li, W., Bahadori, M.T., Liu, Y.: Deep computational phenotyping. In: Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 507–516 (2015). https://doi.org/10.1145/2783258.2783365
Miotto, R., Li, L., Kidd, B.A., Dudley, J.T.: Deep patient: an unsupervised representation to predict the future of patients from the electronic health records. Sci. Rep. 6(1), Article 1 (2016). https://doi.org/10.1038/srep26094
Cheng, Y., Wang, F., Zhang, P., Hu, J.: Risk prediction with electronic health records: a deep learning approach. In: Proceedings of the 2016 SIAM International Conference on Data Mining (SDM), pp. 432–440. Society for Industrial and Applied Mathematics (2016). https://doi.org/10.1137/1.9781611974348.49
Choi, Y., Chiu, C.Y.-I., Sontag, D.: Learning low-dimensional representations of medical concepts. In: AMIA Summits on Translational Science Proceedings, pp. 41–50 (2016)
[1706.03762] Attention Is All You Need (n.d.). https://arxiv.org/abs/1706.03762. Retrieved 5 November 2022
Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding (2019). https://doi.org/10.48550/arXiv.1810.04805
Dosovitskiy, A., et al.: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale (2021). https://doi.org/10.48550/arXiv.2010.11929
Li, Y., et al.: BEHRT: transformer for electronic health records. Sci. Rep. 10(1), Article 1 (2020). https://doi.org/10.1038/s41598-020-62922-y
Song, H., Rajan, D., Thiagarajan, J., Spanias, A.: Attend and diagnose: clinical time series analysis using attention models. In: Proceedings of the AAAI Conference on Artificial Intelligence, 32(1), Article 1 (2018). https://doi.org/10.1609/aaai.v32i1.11635
Choi, E., et al.: Learning the graphical structure of electronic health records with graph convolutional transformer. Proc. AAAI Conf. Artif. Intell. 34(1), 606–613 (2020). https://doi.org/10.1609/aaai.v34i01.5400
Srivastava, N., Salakhutdinov, R.: Learning Representations for Multimodal Data with Deep Belief Nets 8 (n.d.)
Jiang, Y.-G., Wu, Z., Tang, J., Li, Z., Xue, X., Chang, S.-F.: Modeling multimodal clues in a hybrid deep learning framework for video classification. IEEE Trans. Multimedia 20(11), 3137–3147 (2018). https://doi.org/10.1109/TMM.2018.2823900
Cheerla, A., Gevaert, O.: Deep learning with multimodal representation for pancancer prognosis prediction. Bioinformatics 35(14), i446–i454 (2019). https://doi.org/10.1093/bioinformatics/btz342
Vale-Silva, L.A., Rohr, K.: Long-term cancer survival prediction using multimodal deep learning. Sci. Rep. 11(1), Article 1 (2021). https://doi.org/10.1038/s41598-021-92799-4
Meng, Y., Speier, W., Ong, M.K., Arnold, C.W.: Bidirectional representation learning from transformers using multimodal electronic health record data to predict depression. IEEE J. Biomed. Health Inform. 25(8), 3121–3129 (2021). https://doi.org/10.1109/JBHI.2021.3063721
Mai, S., Hu, H., Xing, S.: Modality to Modality Translation: An Adversarial Representation Learning and Graph Fusion Network for Multimodal Fusion (2020). https://doi.org/10.48550/arXiv.1911.07848
De Bois, M., El Yacoubi, M.A., Ammi, M.: Adversarial multi-source transfer learning in healthcare: application to glucose prediction for diabetic people. Comput. Methods Programs Biomed. 199, 105874 (2021). https://doi.org/10.1016/j.cmpb.2020.105874
Zhang, X., et al.: Learning robust patient representations from multi-modal electronic health records: a supervised deep learning approach. In: Proceedings of the 2021 SIAM International Conference on Data Mining (SDM), pp. 585–593 (2020). https://doi.org/10.1137/1.9781611976700.66
Johnson, A.E.W., et al.: MIMIC-III, a freely accessible critical care database. Sci. Data 3(1), Article 1 (2016). https://doi.org/10.1038/sdata.2016.35
Kazemi, S.M., et al.: Time2Vec: Learning a Vector Representation of Time (2019). https://doi.org/10.48550/arXiv.1907.05321
Tipirneni, S., Reddy, C.K.: Self-Supervised Transformer for Sparse and Irregularly Sampled Multivariate Clinical Time-Series (2022). https://doi.org/10.48550/arXiv.2107.14293
Lan, Z., Chen, M., Goodman, S., Gimpel, K., Sharma, P., Soricut, R.: ALBERT: A Lite BERT for Self-supervised Learning of Language Representations (2022). https://doi.org/10.48550/arXiv.1909.11942
Ganin, Y., Lempitsky, V.: Unsupervised Domain Adaptation by Backpropagation (2015). http://arxiv.org/abs/1409.7495
NVIDIA Tesla V100|NVIDIA (n.d.): https://www.nvidia.com/en-gb/data-center/tesla-v100/. Retrieved 21 November 2022
HPC – NUS Information Technology (n.d.): https://nusit.nus.edu.sg/hpc/. Retrieved November 21, 2022
Li, L., et al.: A System for Massively Parallel Hyperparameter Tuning (2020). https://doi.org/10.48550/arXiv.1810.05934
Liaw, R., Liang, E., Nishihara, R., Moritz, P., Gonzalez, J.E., Stoica, I.: Tune: A Research Platform for Distributed Model Selection and Training (2018). http://arxiv.org/abs/1807.05118
Pollard, T.J., Johnson, A.E.W., Raffa, J.D., Celi, L.A., Mark, R.G., Badawi, O.: The eICU Collaborative Research Database, a freely available multi-center database for critical care research. Sci. Data 5(1), Article 1 (2018). https://doi.org/10.1038/sdata.2018.178
Alistair, J., Lucas, B., Tom, P., Steven, H., Leo Anthony, C., Roger, M.: MIMIC-IV (2.1). PhysioNet (n.d.). https://doi.org/10.13026/RRGF-XW32
Multimodal Data Fusion Based on Mutual Information. IEEE Journals & Magazine. IEEE Xplore (n.d.). https://ieeexplore.ieee.org/document/6095545. Retrieved November 21, 2022
Mai, S., Hu, H., Xing, S.: Modality to Modality Translation: An Adversarial Representation Learning and Graph Fusion Network for Multimodal Fusion (2019). http://arxiv.org/abs/1911.07848
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Shankar, B., Hargreaves, C.A. (2023). Adversarial Learning for Improved Patient Representations. In: Frasson, C., Mylonas, P., Troussas, C. (eds) Augmented Intelligence and Intelligent Tutoring Systems. ITS 2023. Lecture Notes in Computer Science, vol 13891. Springer, Cham. https://doi.org/10.1007/978-3-031-32883-1_42
Download citation
DOI: https://doi.org/10.1007/978-3-031-32883-1_42
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-32882-4
Online ISBN: 978-3-031-32883-1
eBook Packages: Computer ScienceComputer Science (R0)