Abstract
Facial Expression Recognition (FER) has received extensive attention in recent years. Due to the strong similarity between expressions, it is urgent to distinguish them meticulously in a finer-grained manner. In this paper, we propose a method, named AU-oriented Expression Decomposition Learning (AEDL), which aims to decouple expressions into Action Units (AUs) and focuses on subtle facial differences. In particular, AEDL comprises two branches: the AU Auxiliary (AUA) branch and the FER branch. For the former, the generic knowledge of dependencies among AUs is leveraged to supervise AU predictions which are then transformed into new expression predictions with a learnable matrix modeled by the relationship between AUs and expressions. For the latter, fusion features are employed to compensate for the minority classes to ensure adequate feature learning. FER predictions are guided by the AUA branch, mining detailed distinctions between expressions. Importantly, the proposed method is independent of the backbone network and brings no extra burden on inference. We conduct experiments on popular in-the-wild datasets and achieve leading performance, proving the effectiveness of the proposed AEDL.
Supported by the National Natural Science Foundation of China (Grant # 62071216, 62231002 and U1936202.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Barsoum, E., Zhang, C., Ferrer, C.C., Zhang, Z.: Training deep networks for facial expression recognition with crowd-sourced label distribution. In: Proceedings of the 18th ACM International Conference on Multimodal Interaction, pp. 279ā283 (2016)
Chen, D., Mei, J.P., Wang, C., Feng, Y., Chen, C.: Online knowledge distillation with diverse peers. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 3430ā3437 (2020)
Cui, Z., Song, T., Wang, Y., Ji, Q.: Knowledge augmented deep neural networks for joint facial expression and action unit recognition. Adv. Neural. Inf. Process. Syst. 33, 14338ā14349 (2020)
Ekman, P., Friesen, W.V.: Facial action coding system. Environ. Psychol. Nonverbal Behav. (1978)
Farzaneh, A.H., Qi, X.: Facial expression recognition in the wild via deep attentive center loss. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 2402ā2411 (2021)
Gu, Y., Yan, H., Zhang, X., Wang, Y., Ji, Y., Ren, F.: Towards facial expression recognition in the wild via noise-tolerant network. IEEE Trans. Circ. Syst. Video Technol. (2022)
Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)
Jiang, J., Deng, W.: Disentangling identity and pose for facial expression recognition. IEEE Trans. Affect. Comput. 13(4), 1868ā1878 (2022)
Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. Comput. Sci. (2014)
Li, S., Deng, W., Du, J.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2584ā2593. IEEE (2017)
Lin, Z., She, J., Shen, Q.: Real emotion seeker: recalibrating annotation for facial expression recognition. Multimedia Syst. 29(1), 139ā151 (2023)
Lucey, P., Cohn, J.F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The extended Cohn-Kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-Workshops, pp. 94ā101. IEEE (2010)
Mollahosseini, A., Hasani, B., Mahoor, M.H.: AffectNet: a database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affect. Comput. 10(1), 18ā31 (2017)
Pu, T., Chen, T., Xie, Y., Wu, H., Lin, L.: Au-expression knowledge constrained representation learning for facial expression recognition. In: 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 11154ā11161. IEEE (2021)
Ruan, D., Yan, Y., Lai, S., Chai, Z., Shen, C., Wang, H.: Feature decomposition and reconstruction learning for effective facial expression recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7660ā7669 (2021)
She, J., Hu, Y., Shi, H., Wang, J., Shen, Q., Mei, T.: Dive into ambiguity: latent distribution mining and pairwise uncertainty estimation for facial expression recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6248ā6257 (2021)
Valstar, M., Pantic, M.: Induced disgust, happiness and surprise: an addition to the mmi facial expression database. In: Proceedings of the 3rd International Workshop on EMOTION (satellite of LREC): Corpora for Research on Emotion and Affect, p. 65. Paris, France (2010)
Wang, K., Peng, X., Yang, J., Lu, S., Qiao, Y.: Suppressing uncertainties for large-scale facial expression recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6897ā6906 (2020)
Xue, F., Wang, Q., Guo, G.: Transfer: learning relation-aware facial expression representations with transformers. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3601ā3610 (2021)
Yang, J., Lv, Z., Kuang, K., Yang, S., Xiao, L., Tang, Q.: RASN: using attention and sharing affinity features to address sample imbalance in facial expression recognition. IEEE Access 10, 103264ā103274 (2022)
Zeng, D., Lin, Z., Yan, X., Liu, Y., Wang, F., Tang, B.: Face2Exp: combating data biases for facial expression recognition. In: 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 20259ā20268 (2022). https://doi.org/10.1109/CVPR52688.2022.01965
Zhang, X., et al.: BP4D-spontaneous: a high-resolution spontaneous 3D dynamic facial expression database. Image Vis. Comput. 32(10), 692ā706 (2014)
Zhang, Y., Dong, W., Hu, B.G., Ji, Q.: Classifier learning with prior probabilities for facial action unit recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5108ā5116 (2018)
Zhao, G., Huang, X., Taini, M., Li, S.Z., PietikƤInen, M.: Facial expression recognition from near-infrared videos. Image Vis. Comput. 29(9), 607ā619 (2011)
Zhao, K., Chu, W.S., De la Torre, F., Cohn, J.F., Zhang, H.: Joint patch and multi-label learning for facial action unit detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2207ā2216 (2015)
Zhao, Z., Liu, Q., Wang, S.: Learning deep global multi-scale and local attention features for facial expression recognition in the wild. IEEE Trans. Image Process. 30, 6544ā6556 (2021)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
Ā© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Lin, Z., She, J., Shen, Q. (2024). AU-Oriented Expression Decomposition Learning for Facial Expression Recognition. In: Liu, Q., et al. Pattern Recognition and Computer Vision. PRCV 2023. Lecture Notes in Computer Science, vol 14429. Springer, Singapore. https://doi.org/10.1007/978-981-99-8469-5_21
Download citation
DOI: https://doi.org/10.1007/978-981-99-8469-5_21
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-8468-8
Online ISBN: 978-981-99-8469-5
eBook Packages: Computer ScienceComputer Science (R0)