Abstract
Decomposing the spatial/temporal functional brain networks from 4D functional magnetic resonance imaging (fMRI) data has attracted extensive attention. Among all these efforts, deep neural network-based methods have shown significant advantages due to their powerful hierarchical representation ability. However, the network architectures of those deep learning models are manually crafted, which is time consuming and non-optimal. This paper presents a novel graph representation neural architecture search (GR-NAS) method based on graph representation to optimize the vanilla RNN cell structure for decomposing spatial/temporal brain networks. The core idea is to embed the discrete search space of the RNN cell into a continuous domain that preserves the topological information. After that, popular search algorithms, e.g., reinforcement learning (RL) and Bayesian optimization (BO), can be employed to find the optimal architecture in this continuous space. The proposed method was evaluated on the Human Connectome Project (HCP) task fMRI datasets. Extensive experiments demonstrated the superiority of the proposed model in brain network decomposition both spatially and temporally. To our best knowledge, the proposed model is among the early efforts using NAS strategy to optimally decompose spatial/temporal functional brain networks from fMRI data.
H. Dai and Q. Li - Co-first authors.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Woolrich, M.W., Jenkinson, M., Brady, J.M., Smith, S.M.: Fully Bayesian spatio-temporal modeling of fMRI data. IEEE Trans. Med. Imaging 23(2), 213–231 (2004)
Zhao, Y., Dai, H., Zhang, W., Ge, F., Liu, T. Two-stage spatial temporal deep learning framework for functional brain network modeling. In: 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), pp. 1576–1580 (2019)
Hu, X., et al.: Latent source mining in FMRI via restricted Boltzmann machine. Hum. Brain Mapp. 39(6), 2368–2380 (2018)
Huang, H., et al.: Modeling task fMRI data via deep convolutional autoencoder. IEEE Trans. Med. Imaging 37(7), 1551–1561 (2017)
Zhang, W., et al.: Hierarchical organization of functional brain networks revealed by hybrid spatiotemporal deep learning. Brain Connect. 10(2), 72–82 (2020)
Zhang, W., Zhao, L., Li, Q., Zhao, S., Dong, Q., Jiang, X., et al.: Identify hierarchical structures from task-based fMRI data via hybrid spatiotemporal neural architecture search net. In: International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI 2019), pp. 745–753 (2019)
Wang, H., et al.: Recognizing brain states using deep sparse recurrent neural network. IEEE Trans. Med. Imaging 38(4), 1058–1068 (2018)
Li, Q., Dong, Q., Ge, F., Qiang, N., Zhao, Y., Wang, H., et al.: Simultaneous spatial-temporal decomposition of connectome-scale brain networks by deep sparse recurrent auto-encoders. In: International Conference on Information Processing in Medical Imaging (IPMI 2019), pp. 579–591 (2019)
Li, Q., Zhang, W., Lv, J., Wu, X., Liu, T.: Neural architecture search for optimization of spatial-temporal brain network decomposition. In: International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI 2020), pp. 377–386 (2020)
Liu, H., Simonyan, K., Yang, Y.: Darts: differentiable architecture search. arXiv preprint arXiv:1806.09055 (2018)
Yan, S., Zheng, Y., Ao, W., Zeng, X., Zhang, M.: Does unsupervised architecture representation learning help neural architecture search?. arXiv preprint arXiv:2006.06936 (2020)
Li, Q., Wu, X., Liu, T.: Differentiable neural architecture search for optimal spatial/temporal brain function network decomposition. Med. Image Anal. 69, 101974 (2021)
Xu, K., Hu, W., Leskovec, J., Jegelka, S.: How powerful are graph neural networks?. arXiv preprint arXiv:1810.00826 (2018)
Zoph, B., Le, Q.V. Neural architecture search with reinforcement learning. arXiv preprint arXiv:1611.01578 (2016)
Williams, R.J.: Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach. Learn. 8(3–4), 229–256 (1992)
White, C., Neiswanger, W., Savani, Y.: Bananas: Bayesian optimization with neural architectures for neural architecture search. arXiv preprint arXiv:1910.11858 (2019)
Gardner, M.W., Dorling, S.R.: Artificial neural networks (the multilayer perceptron)—a review of applications in the atmospheric sciences. Atmos. Environ. 32(14–15), 2627–2636 (1998)
Zou, H., Hastie, T.: Regularization and variable selection via the elastic net. J. R. Stat. Soc. Series B Stat. Methodol. 67(2), 301–320 (2005)
Dice, L.R.: Measures of the amount of ecologic association between species. Ecology 26(3), 297–302 (1945)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 Springer Nature Switzerland AG
About this paper
Cite this paper
Dai, H. et al. (2022). Graph Representation Neural Architecture Search for Optimal Spatial/Temporal Functional Brain Network Decomposition. In: Lian, C., Cao, X., Rekik, I., Xu, X., Cui, Z. (eds) Machine Learning in Medical Imaging. MLMI 2022. Lecture Notes in Computer Science, vol 13583. Springer, Cham. https://doi.org/10.1007/978-3-031-21014-3_29
Download citation
DOI: https://doi.org/10.1007/978-3-031-21014-3_29
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-21013-6
Online ISBN: 978-3-031-21014-3
eBook Packages: Computer ScienceComputer Science (R0)