skip to main content
10.1145/3318299.3318347acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlcConference Proceedingsconference-collections
research-article

Bimodal Emotion Recognition Based on Convolutional Neural Network

Authors Info & Claims
Published:22 February 2019Publication History

ABSTRACT

Computer emotion recognition plays an important role in the field of artificial intelligence and is a key technology to realize human-machine interaction. Aiming at a cross-modal fusion problem of two nonlinear features of facial expression image and speech emotion, a bimodal fusion emotion recognition model (D-CNN) based on convolutional neural network is proposed. Firstly, a fine-grained feature extraction method based on convolutional neural network is proposed. Secondly, in order to obtain joint features representation, a feature fusion method based on the fine-grained features of bimodal is proposed. Finally, in order to verify the performance of the D-CNN model, experiments were conducted on the open source dataset eNTERFACE'05. The experimental results show that the multi-modal emotion recognition model D-CNN is more than 10% higher than the single emotion recognition model of speech and facial expression respectively. In addition, compared with the other commonly used bimodal emotion recognition methods(such as universal background model), the recognition rete of D-CNN is increased by 5%.

References

  1. Rajesh K M, Naveenkumar M. A robust method for face recognition and face emotion detection system using support vector machines{C}// 016 International Conference on Electrical, Electronics, Communication, Computer and Optimization Techniques. IEEE, 2017:1--5.Google ScholarGoogle Scholar
  2. Verma, R., & Dabbagh, M.-Y. (2013). FAST FACIAL EXPRESSION RECOGNITION BASED ON LOCAL BINARY PATTERNS. In 26th IEEE Canadian Conference Of Electrical And Computer Engineering (CCECE) (pp. 1--4).Google ScholarGoogle Scholar
  3. Hasani, B., & Mahoor, M. H. (n.d.). Facial Expression Recognition Using Enhanced Deep 3D Convolutional Neural Networks, 30--40.Google ScholarGoogle Scholar
  4. Barsoum, E., Zhang, C., Ferrer, C. C., & Zhang, Z. (2016). Training Deep Networks for Facial Expression Recognition with Crowd-Sourced Label Distribution, 279--283. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Nema, B. M., & Abdul-kareem, A. A. (2017). Preprocessing Signal for Speech Emotion Recognition, (November).Google ScholarGoogle Scholar
  6. Chaudhari, P. A. A., Pund, M. A., Bamnote, G. R., & Pattalwar, P. S. V. (n.d.). Need of Boosted GMM in Speech Emotion Recognition System Implemented Using Gaussian Mixture Model, 3(1), 1--13.Google ScholarGoogle Scholar
  7. Lim, W., Jang, D., & Lee, T. (2016). Speech emotion recognition using convolutional and Recurrent Neural Networks. 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), 1--4.Google ScholarGoogle ScholarCross RefCross Ref
  8. Huang, Z., Dong, M., Mao, Q., & Zhan, Y. (2014). Speech Emotion Recognition Using CNN. Proceedings of the ACM International Conference on Multimedia - MM '14, 801--804. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Jirayucharoensak, S., Pan-Ngum, S., & Israsena, P. (2014). EEG-Based Emotion Recognition Using Deep Learning Network with Principal Component Based Covariate Shift Adaptation. Scientific World Journal, 2014.Google ScholarGoogle ScholarCross RefCross Ref
  10. Ajay S G, Pravena D, Govind D, et al. Exploring the Significance of Low Frequency Regions in Electroglottographic Signals for Emotion Recognition{J}. 2017.Google ScholarGoogle Scholar
  11. CHEN Peng-zhan, ZHANG Xin, XU Fang-ping.Two-modal emotion recognition based on speech signal and text information{J}. Journal of East China Jiaotong University, 2017, 34(02):100--104.Google ScholarGoogle Scholar
  12. Lee, C. W., Song, K. Y., Jeong, J., & Choi, W. Y. (2018). Convolutional Attention Networks for Multimodal Emotion Recognition from Speech and Text Data, 1--8. https://doi.org/arXiv:1805.06606v1Google ScholarGoogle Scholar
  13. Paper, R. (2013). Towards Efficient Multi-Modal Emotion Recognition, 1--10.Google ScholarGoogle Scholar
  14. Mansoorizadeh, M., & Charkari, N. M. (2010). Multimodal information fusion application to human emotion recognition from face and speech. Multimedia Tools and Applications. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Tzirakis, P., Trigeorgis, G., Nicolaou, M. A., Schuller, B., & Zafeiriou, S. (2016). End-to-End Multimodal Emotion Recognition using Deep Neural Networks, 14(8), 1--9.Google ScholarGoogle Scholar
  16. Poria, S., Chaturvedi, I., & Cambria, E. (2016). Convolutional MKL Based Multimodal Emotion Recognition and Sentiment Analysis.Google ScholarGoogle Scholar
  17. Yan Jingjie, Lu Guanming, Li Haibo, Wang Shanshan. Two-modal emotion recognition based on facial expression and speech{J}. Journal of Nanjing University of Posts and Telecommunications (Natural Science Edition), 2018, 38(01):60--65.Google ScholarGoogle Scholar

Index Terms

  1. Bimodal Emotion Recognition Based on Convolutional Neural Network

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      ICMLC '19: Proceedings of the 2019 11th International Conference on Machine Learning and Computing
      February 2019
      563 pages
      ISBN:9781450366007
      DOI:10.1145/3318299

      Copyright © 2019 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 22 February 2019

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader