Abstract
Brain-computer interface (BCI), a direct communication system between the human brain and external environment, can provide assistance for people with disabilities. Vigilance is an important cognitive state and has a close influence on the performance of users in BCI systems. In this study, a four-target BCI system for cursor control was built based on steady-state visual evoked potential (SSVEP) and twelve subjects were recruited and carried out two long-term BCI experimental sessions, which consisted of two SSVEP-based cursor-control tasks. During each session, electroencephalogram (EEG) signals were recorded. Based on the labeled EEG data of the source domain (previous session) and a small amount of unlabeled EEG data of the target domain (new session), we developed a fine-grained domain adaptation network (FGDAN) for cross-session vigilance estimation in BCI tasks. In the FGDAN model, the graph convolution network (GCN) was built to extract deep features of EEG. The fined-grained feature alignment was proposed to highlight the importance of the different channels figured out by the attention weights mechanism and aligns the feature distributions between source and target domains at the channel level. The experimental results demonstrate that the proposed FGDAN achieved a better performance than the compared methods and indicate the feasibility and effectiveness of our methods for cross-session vigilance estimation of BCI users.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Oken, B.S., Salinsky, M.C., Elsas, S.: Vigilance, alertness, or sustained attention: physiological basis and measurement. Clin. Neurophysiol. 117(9), 1885–1901 (2006)
Zheng, W., et al.: Vigilance estimation using a wearable EOG device in real driving environment. IEEE Trans. Intell. Transp. Syst. 21(1), 170–184 (2020)
Sauvet, F., et al.: In-flight automatic detection of vigilance states using a single EEG channel. IEEE Trans. Biomed. Eng. 61(12), 2840–2847 (2014)
Wolpaw, J., et al.: Brain-computer interfaces for communication and control. Clin. Neurophysiol. 113(6), 767–791 (2002)
Wang, K., et al.: Vigilance estimating in SSVEP-based BCI using multimodal signals. In: 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), pp. 5974–5978 (2021)
Du, R., Liu, R., Wu, T., Lu, B.: Online vigilance analysis combining video and electrooculography features. In: 2012 International Conference on Neural Information Processing (ICONIP), pp. 447–454 (2012)
Krajewski, J., Batliner, A., Golz, M.: Acoustic sleepiness detection: framework and validation of a speech-adapted pattern recognition approach. Behav. Res. Methods 41(3), 795–804 (2009)
Shi, L., Jiao, Y., Lu, B.: Differential entropy feature for EEG-based vigilance estimation. In: 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pp. 6627–6630 (2013)
Shi, L., Lu, B.: EEG-based vigilance estimation using extreme learning machines. Neurocomputing 102, 135–143 (2013)
Zheng, W., Lu, B.: A multimodal approach to estimating vigilance using EEG and forehead EOG. J. Neural Eng. 14, 026017 (2017)
Ko, W., Oh, K., Jeon, E., Suk, H.: VIGNet: a deep convolutional neural network for EEG-based driver vigilance estimation. In: 2020 8th International Winter Conference on Brain-Computer Interface (BCI), pp. 1–3 (2020)
Khessiba, S., Blaiech, A.G., Khalifa, K.B., Abdallah, A.B., Bedoui, M.H.: Innovative deep learning models for EEG-based vigilance detection. Neural Comput. Appl. 33, 6921–6937 (2020)
Zhang, G., Etemad, A.: Capsule attention for multimodal EEG-EOG representation learning with application to driver vigilance estimation. IEEE Trans. Neur. Syst. Rehabil. 29, 1138–1149 (2021)
Jayaram, V., Alamgir, M., Altun, Y., Scholkopf, B., Grosse-Wentrup, M.: Transfer learning in brain-computer interfaces. IEEE Comput. Intell. M. 11(1), 20–31 (2016)
Li, H., Zheng, W., Lu, B.: Multimodal vigilance estimation with adversarial domain adaptation networks. In: 2018 International Joint Conference on Neural Networks (IJCNN), pp. 1–6 (2018)
Luo, Y., Lu, B.: Wasserstein-distance-based multi-source adversarial domain adaptation for emotion recognition and vigilance estimation. In: 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 1424–1428 (2021)
Manyakov, N.V., Chumerin, N., Robben, A., Combaz, A., Van Vliet, M., Van Hulle, M.M.: Sampled sinusoidal stimulation profile and multichannel fuzzy logic classification for monitor-based phase-coded SSVEP brain-computer interfacing. J. Neural Eng. 10, 036011 (2013)
Chen, X., Wang, Y., Gao, S., Jung, T.-P., Gao, X.: Filter bank canonical correlation analysis for implementing a high-speed SSVEP-based brain-computer interface. J. Neural Eng. 12, 046008 (2015)
Gomez-Herrero, G., et al.: Automatic removal of ocular artifacts in the EEG without an EOG reference channel. In: Proceedings of the 7th Nordic Signal Processing Symposium, pp. 130–133 (2006)
Jia, Z., et al.: GraphSleepNet: adaptive spatial-temporal graph convolutional networks for sleep stage classification. In: Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI 2020), pp. 1324–1330 (2021)
Defferrard, M., Bresson, X., Vandergheynst, P.: Convolutional neural networks on graphs with fast localized spectral filtering. In: Proceedings of the 30th International Conference on Neural Information Processing Systems, pp. 3844–3852 (2016)
Gretton, A., et al.: A kernel two-sample test. J. Mach. Learn. Res. 13(25), 723–773 (2012)
Hu, J., Shen, L., Albanie, S., Sun, G., Wu, E.: Squeeze-and-excitation networks. IEEE Trans. Pattern Anal. Mach. Intell. 42(8), 2011–2023 (2020)
Shen, F., Dai, G., Lin, G., Zhang, J., Zeng, H.: EEG-based emotion recognition using 4D convolutional recurrent neural network. Cogn. Neurodyn. 14, 815–828 (2020)
Gao, Z., Wang, X., Yang, Y., Li, Y., Ma, K., Chen, G.: A channel-fused dense convolutional network for EEG-based emotion recognition. IEEE T. Cogn. Dev. Syst. 13, 945–954 (2021)
Tzeng, E., et al.: Deep domain confusion: maximizing for domain invariance. arXiv preprint arXiv:1412.3474 (2014)
Long, M., Cao, Y., Wang, J., Jordan, M.: Learning transferable features with deep adaptation networks. In: Proceedings of the 32nd International Conference on Machine Learning, pp. 97–105 (2015)
Ganin, Y., et al.: Domain-adversarial training of neural networks. J. Mach. Learn. Res. 17(59), 1–35 (2016)
Chen, X., Wang, S., Wang, J., Long, M.: Representation subspace distance for domain adaptation regression. In: Proceedings of the 38th International Conference on Machine Learning, pp. 1749–1759 (2021)
Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9, 2579–2605 (2008)
Acknowledgements
This work was supported by the Beijing Natural Science Foundation [grant numbers 7222311 and J210010], the National Natural Science Foundation of China [grant numbers U21A20388 and 62206285] and China Postdoctoral Science Foundation (grant number 2021M703490).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Wang, K. et al. (2024). A Fine-Grained Domain Adaptation Method for Cross-Session Vigilance Estimation in SSVEP-Based BCI. In: Luo, B., Cheng, L., Wu, ZG., Li, H., Li, C. (eds) Neural Information Processing. ICONIP 2023. Lecture Notes in Computer Science, vol 14449. Springer, Singapore. https://doi.org/10.1007/978-981-99-8067-3_6
Download citation
DOI: https://doi.org/10.1007/978-981-99-8067-3_6
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-8066-6
Online ISBN: 978-981-99-8067-3
eBook Packages: Computer ScienceComputer Science (R0)