Abstract
Recognition of users’ facial expressions and reflecting them on the face of the user’s virtual avatar is a key technology for realizing immersive virtual reality (VR)-based metaverse applications. As a method to realize this technology, a facial electromyogram (fEMG)-based facial expression recognition (FER) system, with the fEMG electrodes being attached on the pad of a VR headset, has recently been proposed. However, the performance of such FER systems has severely deteriorated when the locations of fEMG electrodes change by the re-wearing of the VR headset, requiring long and tedious calibration sessions every time the user wears the VR headset. In this study, we developed an fEMG-based FER system that is robust against electrode shifts by employing new signal processing techniques: covariate shift adaptation techniques in feature and classifier domains. To verify the feasibility of the proposed method, fEMG data were recorded while participants were making 11 facial expressions repeatedly in four sessions, between which they detached and reattached the fEMG electrodes on their faces. Our experiments showed that classification accuracy dropped from 88 to 79% by the change of the electrode locations when the proposed method was not applied, whereas the accuracy was significantly improved up to 86% when the proposed covariate shift adaptation method was employed. It is expected that the proposed method would contribute to enhancing the practicality of the fEMG-based FER, promoting the practical application of the fEMG-based FER to VR-based metaverse applications.







Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Availability of data and materials
Accessible. We uploaded the dataset (bdf format), which is available at https://figshare.com/s/bed96c783b4328d4ad1d, with the data description document (doc format).
Code availability
Not applicable.
Notes
References
Arsigny V, Fillard P, Pennec X, Ayache N (2007) Geometric means in a novel vector space structure on symmetric positive-definite matrices. SIAM J Matrix Anal Appl 29:328–347. https://doi.org/10.1137/050637996
Asghari Oskoei M, Hu H (2007) Myoelectric control systems—a survey. Biomed Signal Process Control 2:275–294. https://doi.org/10.1016/J.BSPC.2007.07.009
Barachant A, Bonnet S, Congedo M, Jutten C (2013) Classification of covariance matrices using a riemannian-based kernel for BCI applications. Neurocomputing 112:172–178. https://doi.org/10.1016/j.neucom.2012.12.039
Barachant A, Bonnet S, Congedo M, Jutten C (2010) Riemannian Geometry Applied to BCI Classification. In: Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). pp 629–636
Bouveyron C, Brunet C (2012) Probabilistic fisher discriminant analysis: a robust and flexible alternative to fisher discriminant analysis. Neurocomputing 90:12–22. https://doi.org/10.1016/j.neucom.2011.11.027
Caserman P, Garcia-Agundez A, Konrad R et al (2019) Real-time body tracking in virtual reality using a Vive tracker. Virtual Real 23:155–168. https://doi.org/10.1007/s10055-018-0374-z
Cha H-S, Im C-H (2021) Performance enhancement of facial electromyogram-based facial-expression recognition for social virtual reality applications using linear discriminant analysis adaptation. Virtual Real 1:1–14. https://doi.org/10.1007/s10055-021-00575-6
Cha H-S, Choi S-J, Im C-H (2020) Real-time recognition of facial expressions using facial electromyograms recorded around the eyes for social virtual reality applications. IEEE Access 8:62065–62075. https://doi.org/10.1109/access.2020.2983608
Chen Y, Yang Z, Wang J (2015) Eyebrow emotional expression recognition using surface EMG signals. Neurocomputing 168:871–879. https://doi.org/10.1016/j.neucom.2015.05.037
Driscoll WC (1996) Robustness of the ANOVA and Tukey-Kramer statistkal tests. Comput Ind Eng. https://doi.org/10.1016/0360-8352(96)00127-1
Ekman P (1993) Facial expression and emotion. Am Psychol 48:384–392. https://doi.org/10.1037/0003-066X.48.4.384
Ekman P, Rosenberg EL (2005) What the face revealsbasic and applied studies of spontaneous expression using the facial action coding system (FACS). Oxford University Press
Fatoorechi M, Archer J, Nduka C, et al (2017) Using facial gestures to drive narrative in VR. In: SUI 2017-Proceedings of the 2017 Symposium on Spatial User Interaction. ACM Press, New York, USA, p 152
Förstner W, Moonen B (2003) A metric for covariance matrices. In: Grafarend EW, Krumm FW, Schwarze VS (eds) Geodesy-the challenge of the 3rd millennium. Springer, Berlin Heidelberg, pp 299–309
Fox J, Arena D, Bailenson JN (2009) Virtual reality: a survival guide for the social scientist. J Media Psychol 21:95–113. https://doi.org/10.1027/1864-1105.21.3.95
Gonzalez-Franco M, Steed A, Hoogendyk S, Ofek E (2020) Using facial animation to increase the enfacement illusion and avatar self-identification. IEEE Trans vis Comput Graph 26:2023–2029. https://doi.org/10.1109/TVCG.2020.2973075
Guevara JE, Mogollón H, Pitman NCA et al (2017) Improving the robustness of myoelectric pattern recognition for upper limb prostheses by covariate shift adaptation. IEEE Trans Neural Syst Rehabil Eng 24:27–52. https://doi.org/10.1002/9781119090670.ch2
Hakonen M, Piitulainen H, Visala A (2015) Current state of digital signal processing in myoelectric interfaces and related applications. Biomed Signal Process Control 18:334–359. https://doi.org/10.1016/j.bspc.2015.02.009
Hamedi M, Salleh S-H, Swee TT et al (2011) Surface electromyography-based facial expression recognition in Bi-polar configuration. J Comput Sci 7:1407
Hamedi M, Salleh SH, Ting CM et al (2018) Robust facial expression recognition for MuCI: a comprehensive neuromuscular signal analysis. IEEE Trans Affect Comput 9:102–115. https://doi.org/10.1109/TAFFC.2016.2569098
Hargrove L, Englehart K, Hudgins B (2008) A training strategy to reduce classification degradation due to electrode displacements in pattern recognition based myoelectric control. Biomed Signal Process Control 3:175–180. https://doi.org/10.1016/j.bspc.2007.11.005
Hickson S, Kwatra V, Dufour N et al (2015) Facial performance sensing head-mounted display. ACM Trans Graph 34:47. https://doi.org/10.1145/2766939
Hickson S, Kwatra V, Dufour N, et al (2019) Eyemotion: Classifying facial expressions in VR using eye-tracking cameras. In: IEEE Winter Conference on Applications of Computer Vision. IEEE, pp 1626–1635
Hiraoka K, Hamahira M, Hidai KI et al (2001) Fast algorithm for online linear discriminant analysis. IEICE Trans Fundam Electron Commun Comput Sci E84-A:1431–1440
Htut K-M, Tamaki H, Nakajima A, Shigehara; T (2002) Fast algorithm for updating discriminant functions in linear discriminant analysis. In: Proceedings of IEEK Conferences 2008–2011
Kumar S, Yger F, Lotte F (2019) Towards adaptive classification using riemannian geometry approaches in brain-computer interfaces. In: 7th International Winter Conference on Brain-Computer Interface, BCI 2019
Langner O, Dotsch R, Bijlstra G et al (2010) Presentation and validation of the radboud faces database. Cogn Emot 24:1377–1388. https://doi.org/10.1080/02699930903485076
Lee J, Kim M, Kim J (2020) RoleVR: multi-experience in immersive virtual reality between co-located HMD and non-HMD users. Multimed Tools Appl 79:979–1005. https://doi.org/10.1007/s11042-019-08220-w
Li L, Yu F, Shi D et al (2017) Application of virtual reality technology in clinical medicine. Am J Transl Res 9:3867–3880
Lou J, Wang Y, Nduka C et al (2020) Realistic facial expression reconstruction for VR HMD users. EEE Trans Multimed 22:730–743. https://doi.org/10.1109/TMM.2019.2933338
Ma M, Zheng H (2011) Virtual reality and serious games in healthcare. In: Brahnam S, Jain LC (eds) Studies in computational intelligence. Springer-Verlag, Berlin Heidelberg, pp 169–192
Mavridou I, McGhee JT, Hamedi M, et al (2017) FACETEQ interface demo for emotion expression in VR. In: IEEE Virtual Reality. pp 441–442
Mikropoulos TA, Natsis A (2011) Educational virtual environments: a ten-year review of empirical research (1999–2009). Comput Educ 56:769–780. https://doi.org/10.1016/j.compedu.2010.10.020
Morerio P, Murino V (2017) Correlation alignment by riemannian metric for domain adaptation. arXiv
Morrison DG (1969) On the interpretation of discriminant analysis. J Mark Res 6:156. https://doi.org/10.2307/3149666
Olszewski K, Lim JJ, Saito S, Li H (2016) High-fidelity facial and speech animation for VR HMDs. ACM Trans Graph 35:1–14. https://doi.org/10.1145/2980179.2980252
Ostertagová E, Ostertag O, Kováč J (2014) Methodology and application of the Kruskal-Wallis test. Appl Mech Mater. https://doi.org/10.4028/www.scientific.net/AMM.611.115
Psotka J (1995) Immersive training systems: Virtual reality and education and training. Instr Sci 23:405–431. https://doi.org/10.1007/BF00896880
Sato W, Yoshikawa S (2007) Spontaneous facial mimicry in response to dynamic facial expressions. Cognition 104:1–18. https://doi.org/10.1016/j.cognition.2006.05.001
Saxena V V., Feldt T, Goel M (2014) Augmented telepresence as a tool for immersive simulated dancing in experience and learning. In: Proceedings of the India HCI 2014 Conference on Human Computer Interaction. pp 86–89
Sugiyama M, Krauledat M, Müller KR (2007) Covariate shift adaptation by importance weighted cross validation. J Mach Learn Res
Thies J, Zollhöfer M, Stamminger M et al (2018) FaceVR: real-time gaze-aware facial reenactment in virtual reality. ACM Trans Graph. https://doi.org/10.1145/3182644
Vidaurre C, Kawanabe M, Von Bünau P et al (2011) Toward unsupervised adaptation of LDA for brain-computer interfaces. IEEE Trans Biomed Eng 58:587–597. https://doi.org/10.1109/TBME.2010.2093133
Wang R, Guo H, Davis LS, Dai Q (2012) Covariance discriminative learning: a natural and efficient approach to image set classification. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. pp 2496–2503
Yger F, Berar M, Lotte F (2017) Riemannian approaches in brain-computer interfaces: a review. IEEE Trans Neural Syst Rehabil Eng 25:1753–1762. https://doi.org/10.1109/TNSRE.2016.2627016
Acknowledgements
This work was supported by the Institute of Information & communications Technology Planning & Evaluation (IITP) grants funded by the Korea government (MIST) (Nos. 2017-0-00432 & 2020-0-01373).
Funding
This work was supported by the Institute for Information and Communications Technology Promotion, (Grant Nos. 2017-0-00432, 2020-0-01373).
Author information
Authors and Affiliations
Contributions
H. Cha conducted overall data analyses and wrote a major part of the paper. C. Im provided important insight for the design of the paper and revised the manuscript. All authors listed have contributed considerably to this paper and approved the submitted version.
Corresponding author
Ethics declarations
Conflicts of interest
The authors declare that there are no conflicts of interest regarding the publication of this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix A
Geodesic on a Riemannian manifold is the shortest path between two SPD matrices on a Riemannian manifold (Yger et al. 2017). The Geodesic between \({{\varvec{C}}}_{1}\) and \({{\varvec{C}}}_{2}\) is defined as
where \(c\in \left[0, 1\right]\). Note that output of the \(\gamma \left({{\varvec{C}}}_{1},{{\varvec{C}}}_{2}, c\right)\) is located between \({{\varvec{C}}}_{i}\) and \({{\varvec{C}}}_{2}\) depending on the constant \(c\). For example, \(\gamma \left({{\varvec{C}}}_{1},{{\varvec{C}}}_{2},c\right)\) is \({{\varvec{C}}}_{1}\) if \(c=0\) and \({{\varvec{C}}}_{2}\) if \(c=1\). \(\gamma \left({{\varvec{C}}}_{1},{{\varvec{C}}}_{2},c\right)\) will be placed at a center point between \({{\varvec{C}}}_{1}\) and \({{\varvec{C}}}_{2}\) along the geodesic if \(c=0.5\).
Appendix B
The distance between two SPD matrices (Yger et al. 2017) on the Riemannian manifold can be defined as
where the \(\mathrm{logm}\) is the logarithm of a matrix and \(\Vert \cdot \Vert\) is the Frobenius norm of a matrix. Equation (7) can be easily computed by \({\left[\sum_{i=1}^{n}{\lambda }_{i}\right]}^\frac{1}{2}\) where \({\lambda }_{i}\) s are the real positive eigenvalues of \({{\varvec{C}}}_{1}^{-1}{{\varvec{C}}}_{2}\).
Appendix C
The geometric mean is defined as
where \(C\left(n\right)\) is the set of all \(n\times n\) SPD matrices. Equation (8) is not closed form; therefore, interactive algorithm (Barachant et al. 2010) can be employed instead of this, which is written as follows:

Appendix D
Statistical derivation of LDA classification is as follow. The LDA assumes that data within a class label has the multivariate normal distribution. The probability density function (pdf) that feature vector \({{\varvec{x}}}_{i}\) given that label \({y}_{i}\) is \(k\) can be defined as
where \({{\varvec{\mu}}}_{k}\in {R}^{36}\) is a mean vector of feature vector \({{\varvec{x}}}_{i}\) within the label \(k\) and \(\boldsymbol{\Sigma }\in {R}^{36\times 36}\) is a pooled covariance matrix (PCM). \({{\varvec{\mu}}}_{k}\) and \(\boldsymbol{\Sigma }\) can be estimated as
where \(N\), \({N}_{k}\) and \(K\) are the number of total samples of feature vectors, the number of samples of feature vector within a label \(k\), and the total number of labels, respectively.
The probability that label \(k\) is classified given feature vector \({{\varvec{x}}}_{i}\) can be written using Bayes rules as
\(p\left(y=k\right)\) is the prior probability and can be estimated as
where \(\delta \left(i,j\right)=1\) if \(i=j\) and 0 if \(i\ne j\). Let \(p\left(y=k\right)\) and \(p({\varvec{x}}={{\varvec{x}}}_{{\varvec{i}}}|y=k)\) represent \({\pi }_{k}\) and \({f}_{k}\left({\varvec{x}}\right)\), respectively, then \(p\left(y=k|{\varvec{x}}={{\varvec{x}}}_{{\varvec{i}}}\right)\)~\({\pi }_{k}{f}_{k}({\varvec{x}})\). \({\pi }_{k}{f}_{k}({\varvec{x}})\) is monotonic increment function; \(p\left(y=k|{\varvec{x}}={{\varvec{x}}}_{{\varvec{i}}}\right)\)~\({\pi }_{k}{f}_{k}({\varvec{x}})\)~\(\mathrm{log}({\pi }_{k}{f}_{k}\left({\varvec{x}}\right))\). Let \(\mathrm{log}({\pi }_{k}{f}_{k}\left({\varvec{x}}\right))\) be the decision function \({\varphi }_{k}({\varvec{x}})\), then \({\varphi }_{k}({\varvec{x}})\) can be represented by
Finally, the predicted label \({\widehat{y}}_{j}\) can be estimated with the test data \({{\varvec{x}}}_{j}\) as follows:
Simply put, in the training stage, mean vector \({{\varvec{\mu}}}_{k}\) for every class label (\(y=1, 2, 3, \dots K\)) and \(\boldsymbol{\Sigma }\) were estimated with training dataset using (10) and (11). In the test stage, a test feature vector \({{\varvec{x}}}_{j}\) unseen in training dataset is predicted using (15).
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Cha, HS., Im, CH. Improvement of robustness against electrode shift for facial electromyogram-based facial expression recognition using domain adaptation in VR-based metaverse applications. Virtual Reality 27, 1685–1696 (2023). https://doi.org/10.1007/s10055-023-00761-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10055-023-00761-8