Abstract
This paper presents a strategy to perform an emotional states recognition process by analyzing electroencephalography records; The recognition process were performed by a specific purpose neural network and the experimental criteria for it configuration are presented. Also a novelty electrode discriminant process were applied, which correlates electrodes to Brodmann regions, achieving a data reduction of 29.5 percent. The recognition rates average achieved up to 90.2 percent of recognition rate in the binary case and up to 82.51 percent in a multi-class scheme.
A.R. Aguiñaga—Conacyt and Instituto Tecnológico de Tijuana.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Information processing (CNS), Support (CNS, NES, ANS), Executive (CNS), Action (SNS), Monitor (CNS).
- 2.
(CNS), central nervous system; (NES), neuro-endocrine system; (ANS), autonomic nervous system; (SNS), somatic nervous system. The organismic subsystems are theoretically postulated functional units or networks.
- 3.
Visual activity are related to occipital region (Brodmann region 17), auditory activity are associated to temporal lobe (region 41), the limbic system as the main responsible to emotional processes are located in the temporal parietal areas and body sensations are located in the temporal and parietal regions [25–27].
- 4.
A neural network with up to fifty units per hidden layer and up to filthy hidden layers.
- 5.
Each performed under a 10-fold cross validation.
References
Picard, R.W.: Affective Computing, 1st edn. MIT Press, Cambridge (2000)
Murugappan, M.: Human emotion classification using wavelet transform and knn. In: 2011 International Conference on Pattern Analysis and Intelligent Robotics (ICPAIR), vol. 1, pp. 148–153, June 2011
Murugappan, M., Rizon, M., Nagarajan, R., Yaacob, S., Zunaidi, I., Hazry, D.: Lifting scheme for human emotion recognition using eeg. In: International Symposium on Information Technology, 2008, ITSim 2008, vol. 2, pp. 1–7, August 2008
Koelstra, S., Muhl, C., Soleymani, M., Lee, J.-S., Yazdani, A., Ebrahimi, T., Pun, T., Nijholt, A., Patras, I.: Deap: a database for emotion analysis; using physiological signals. IEEE Trans. Affect. Comput. 3(1), 18–31 (2012)
Sanei, S., Chambers, J.A.: Eeg signal processing. In: Cardiff University (2007)
Lin, Y.-P., Wang, C.-H., Wu, T.-L., Jeng, S.-K., Chen, J.-H.: Multilayer perceptron for eeg signal classification during listening to emotional music. In: TENCON 2007–2007 IEEE Region 10 Conference, pp. 1–3, October 2007
Heraz, A., Razaki, R., Frasson, C.: Using machine learning to predict learner emotional state from brainwaves. In: Seventh IEEE International Conference on Advanced Learning Technologies, 2007, ICALT 2007, pp. 853–857, July 2007
Petrantonakis, P.C., Hadjileontiadis, L.J.: Eeg-based emotion recognition using hybrid filtering and higher order crossings. In: 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops, 2009, ACII 2009, pp. 1–6, September 2009
Takahashi, K.: Remarks on svm-based emotion recognition from multi-modal bio-potential signals. In: 13th IEEE International Workshop on Robot and Human Interactive Communication, 2004, ROMAN 2004, pp. 95–100, September 2004
Li, M., Lu, B.-L.: Emotion classification based on gamma-band eeg. In: Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 2009, EMBC 2009, pp. 1223–1226, September 2009
Rozgic, V., Ananthakrishnan, S., Saleem, S., Kumar, R., Prasad, R.: Ensemble of svm trees for multimodal emotion recognition. In: Signal Information Processing Association Annual Summit and Conference (APSIPA ASC), 2012 Asia-Pacific, pp. 1–4, December 2012
Attabi, Y., Dumouchel, P.: Emotion recognition from speech: woc-nn and class-interaction. In: 2012 11th International Conference on Information Science, Signal Processing and their Applications (ISSPA), pp. 126–131, July 2012
Razak, A.A., Komiya, R., Izani, M., Abidin, Z.: Comparison between fuzzy and nn method for speech emotion recognition. In: Third International Conference on Information Technology and Applications, 2005, ICITA 2005, vol. 1, pp. 297–302, July 2005
Gunler, M.A., Tora, H.: Emotion classification using hidden layer outputs. In: 2012 International Symposium on Innovations in Intelligent Systems and Applications (INISTA), pp. 1–4, July 2012
Bradley, M.M., Cuthbert, B.N., Lang, P.J.: International affective picture system (iaps): Technical manual and affective ratings. In: NIMH Center for the Study of Emotion and Attention, FL: The Center for Research in Psychophysiology, University of Florida (1997)
Bradley, M.M., Cuthbert, B.N., Lang, P.J.: International affective digitized sounds (iads): Stimuli, instruction manual and affective ratings, In: The Center for Research in Psychophysiology, University of Florida (1999)
Yin, L., Wei, X., Sun, Y., Wang, J., Rosato, M.J.: A 3d facial expression database for facial behavior research. In: 7th International Conference on Automatic Face and Gesture Recognition, 2006, FGR 2006, pp. 211–216, April 2006
Moody, G.B., Mark, R.G., Goldberger, A.L.: Physionet: a web-based resource for the study of physiologic signals. Eng. Med. Biol. Mag. IEEE 20(3), 70–75 (2001)
Petridis, S., Martinez, B., Pantic, M.: The MAHNOB laughter database. Image Vis. Comput. 31(2), 186–202 (2013). Affect Analysis In Continuous Input
Scherer, K.R.: What are emotions? and how can they be measured? In: Trends and Developments: Research on Emotions, vol. 44, pp. 695–729. Social Science Information and SAGE Publications (2005)
Friesen, W.V., O’Sullivan, M., Chan, A., Diacoyanni-Tarlatzis, I., Heider, K., Krause, R., LeCompte, W.A., Pitcairn, T., Ricci-Bitti, P.E., Scherer, K., Tomita, M., Tzavaras, A., Ekman, P.: Universals and cultural differences in the judgments of facial expressions of emotion. J. Pers. Soc. Psychol. 4, 712–717 (1987)
Russell, J.A.: A circumplex model of affect. J. Pers. Soc. Psychol. 39, 1161–1178 (1980)
Gerrod Parrott, W.: Emotions in Social Psychology, 1st edn. Psychology Press, Philadelphia (2001)
Ramirez, M.A.L., Aguinaga, A.R., Flores, B.M.R.: Classification model of arousal and valence mental states by eeg signals analysis and brodmann correlations. Int. J. Adv. Comput. Sci. Appl. (IJACSA) 6, 230–238 (2015)
Brodmann, K., Garey, L.J.: Brodmann’s Localisation in the Cerebral Cortex, 3rd edn. Springer Science, USA (2006)
Fox, S.I.: Human Physiology, 7th edn. Mcgraw-Hill (Tx), Boston (2002)
Sherwood, L.: Human Physiology from Cells to System, 8th edn. Brooks/Cole, Cengage Learning, Belmont (2013)
Cecotti, H., Graser, A.: Convolutional neural networks for p300 detection with application to brain-computer interfaces. IEEE Trans. Pattern Anal. Mach. Intell. 33(3), 433–445 (2011)
Pourtois, G., Schwartz, S., Seghier, M.L., Scherer, K.R., Vuilleumier, P., Sander, D., Grandjean, D.: Emotion and attention interactions in social cognition: brain regions involvedin processing anger prosody. NeuroImage Spec. Section Soc. Cogn. Neurosci. 28, 848–858 (2005)
Acknowledgment
To the Instituto Tecnológico de Tijuana and Conacyt for making possible this project.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Aguiñaga, A.R., Ramirez, M.A.L. (2015). Low Complexity Neural Networks to Classify EEG Signals Associated to Emotional Stimuli. In: Bravo, J., Hervás, R., Villarreal, V. (eds) Ambient Intelligence for Health. AmIHEALTH 2015. Lecture Notes in Computer Science(), vol 9456. Springer, Cham. https://doi.org/10.1007/978-3-319-26508-7_18
Download citation
DOI: https://doi.org/10.1007/978-3-319-26508-7_18
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-26507-0
Online ISBN: 978-3-319-26508-7
eBook Packages: Computer ScienceComputer Science (R0)