Fuzzy Multiview Graph Learning on Sparse Electronic Health Records | IEEE Journals & Magazine | IEEE Xplore

Fuzzy Multiview Graph Learning on Sparse Electronic Health Records


Abstract:

Extracting latent disease patterns from electronic health records (EHRs) is a crucial solution for disease analysis, significantly facilitating healthcare decision-making...Show More

Abstract:

Extracting latent disease patterns from electronic health records (EHRs) is a crucial solution for disease analysis, significantly facilitating healthcare decision-making. Multiview learning presents itself as a promising approach that offers a comprehensive exploration of both structured and unstructured EHRs. However, the intrinsic uncertainty among disease features presents a significant challenge for multiview feature alignment. Besides, the sparsity of real-world EHRs also exacerbates the difficulty of feature alignment. To address these challenges, we introduce a novel fuzzy multiview graph learning framework named FuzzyMVG, which is designed for mitigating the impacts of uncertainty in disease features derived from sparse EHRs. First, we utilize auxiliary information from sparse EHRs to construct a multiview EHR graph using the structured and unstructured records. Then, for efficient feature alignment, we specially design the fuzzy logic-enhanced graph convolutional networks to obtain the fuzzy representations of time-invariant node features. Thereby, we implement a random walk strategy and long short-term memory networks to capture the distinct features of static and dynamic nodes, respectively. Extensive experiments have been conducted on the real-world MIMIC III dataset to validate the effectiveness of FuzzyMVG. Results in the diagnosis prediction task demonstrate that FuzzyMVG outperforms other state-of-the-art baselines.
Published in: IEEE Transactions on Fuzzy Systems ( Volume: 32, Issue: 10, October 2024)
Page(s): 5520 - 5532
Date of Publication: 26 June 2024

ISSN Information:

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.