Abstract
The goal of this paper is to evaluate the human perception regarding geometric features, personalities and emotions in avatars. To achieve this, we used a dataset that contains pedestrian tracking files captured in spontaneous videos which are visualized as identical virtual human beings. The main objective is to focus on individuals motion, not having the distraction of other features. In addition to tracking files containing pedestrian positions, the dataset also contains emotion and personality data for each pedestrian, detected through computer vision and pattern recognition techniques. We are interested in evaluating whether participants can perceive geometric features such as density levels, distances, angular variations and speeds, as well as cultural features (emotions and personality traits) in short video sequences (scenes), when pedestrians are represented by avatars. With this aim in mind, we propose two questions to be answered through this analysis: i) “Can people perceive geometric features in avatars?"; and ii) “Can people perceive differences regarding personalities and emotions in virtual humans without body and facial expressions?". Regarding the participants, 73 people volunteered for the experiment in order to answer the two mentioned questions. Results indicate that, even without explaining to the participants the concepts of cultural features and how they were calculated (considering the geometric features), in most cases the participants perceived the personality and emotion expressed by avatars, even without faces and body expressions.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
(Available at: http://www.inf.pucrs.br/vhlab).
- 2.
Social space is related to 3.6 m [4].
- 3.
Unity3D is available at https://unity3d.com/.
References
Alameda-Pineda, X., Ricci, E., Sebe, N.: Multimodal Behavior Analysis in the Wild: Advances and Challenges. Elsevier Science, London (2018)
Alexander, R.M.: A model of bipedal locomotion on compliant legs. Phil. Trans. R. Soc. Lond. B 338(1284), 189–198 (1992)
Araujo, V., Migon Favaretto, R., Knob, P., Raupp Musse, S., Vilanova, F., Brandelli Costa, A.: How much do you perceive this?: an analysis on perceptions of geometric features, personalities and emotions in virtual humans. In: Proceedings of the 19th ACM International Conference on Intelligent Virtual Agents, pp. 179–181. ACM (2019)
Calvin, S., Hall, G.L., Campbell, J.B.: Theories Of Personality, 4th edn. John Wiley & Sons, New Jersey (1998)
Chattaraj, U., Seyfried, A., Chakroborty, P.: Comparison of pedestrian fundamental diagram across cultures. Adv. Complex Syst. 12(03), 393–405 (2009)
Costa, P., McCrae, R.: Revised NEO Personality Inventory (NEO PI-R) and NEO Five-Factor Inventory (NEO-FFI). PAR (1992)
Durupınar, F., Güdükbay, U., Aman, A., Badler, N.I.: Psychological parameters for crowd simulation: from audiences to mobs. IEEE TVCG 22(9), 2145–2159 (2016)
Durupinar, F., Kapadia, M., Deutsch, S., Neff, M., Badler, N.I.: Perform: perceptual approach for adding ocean personality to human motion using laban movement analysis. ACM Trans. Graph. (TOG) 36(1), 1–16 (2016)
Ekman, P., Friesen, W.V.: Constants across cultures in the face and emotion. JPSP 17(2), 124 (1971)
Favaretto, R.M., Dihl, L., Barreto, R., Musse, S.R.: Using group behaviors to detect hofstede cultural dimensions. In: ICIP, pp. 2936–2940. IEEE (2016)
Favaretto, R.M., Dihl, L., Musse, S.R.: Detecting crowd features in video sequences. In: Proceedings of Conference on Graphics, Patterns and Images (SIBGRAPI), pp. 201–208. IEEE Computer Society, São José dos Campos (2016)
Favaretto, R.M., Dihl, L., Musse, S.R., Vilanova, F., Costa, A.B.: Using big five personality model to detect cultural aspects in crowds. In: Graphics, Patterns and Images (SIBGRAPI), pp. 223–229. IEEE (2017)
Favaretto, R.M., Knob, P., Musse, S.R., Vilanova, F., Costa, Â.B.: Detecting personality and emotion traits in crowds from video sequences. Mach. Vision Appl. 30(5), 999–1012 (2018). https://doi.org/10.1007/s00138-018-0979-y
Favaretto, R.M., Musse, S.R., Costa, A.B.: Emotion, personality, and cultural aspects in crowds. In: Emotion, Personality and Cultural Aspects in Crowds, pp. 23–33. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-22078-5_3
Goldenberg, A., Weisz, E., Sweeny, T.D., Cikara, M., Gross, J.J.: The crowd-emotion-amplification effect. Psychol. Sci. 32(3), 437–450 (2021)
Gorbova, J., Lüsi, I., Litvin, A., Anbarjafari, G.: Automated screening of job candidate based on multimodal video processing. In: CVPRW (2017)
Hofstede, G.: Culture’s Consequences: Comparing Values, Behaviors, Institutions and Organizations Across Nations. Sage publications, Thousand Oaks (2001)
Hofstede, G.: Dimensionalizing cultures: the hofstede model in context. Online Read. Psychol. Cult. 2(1), 8 (2011)
Jacques Junior, J., Musse, S.R., Jung, C.: Crowd analysis using computer vision techniques. IEEE Signal Process. Mag. 27, 66–77 (2010)
Knob, P., de Andrade Araujo, V.F., Favaretto, R.M., Musse, S.R.: Visualization of interactions in crowd simulation and video sequences. In: Brazilian Symposium on Computer Games and Digital Entertainment (2018)
Lala, D., Thovuttikul, S., Nishida, T.: Towards a virtual environment for capturing behavior in cultural crowds. In: 6th ICDIM, pp. 310–315 (2011)
Lamer, S.A., Sweeny, T.D., Dyer, M.L., Weisbuch, M.: Rapid visual perception of interracial crowds: racial category learning from emotional segregation. J. Exp. Psychol. Gener. 147(5), 683 (2018)
McDonnell, R., Larkin, M., Dobbyn, S., Collins, S., O’Sullivan, C.: Clone attack! perception of crowd variety. In: ACM SIGGRAPH 2008 papers, pp. 1–8. ACM (2008)
Saifi, L., Boubetra, A., Nouioua, F.: An approach for emotions and behavior modeling in a crowd in the presence of rare events. AB 24(6), 428–445 (2016)
Yang, F., Shabo, J., Qureshi, A., Peters, C.: Do you see groups?: the impact of crowd density and viewpoint on the perception of groups. In: IVA, pp. 313–318, ACM (2018)
Zhao, J., Gou, L., Wang, F., Zhou, M.: Pearl: an interactive visual analytic tool for understanding personal emotion style derived from social media. In: VAST, pp. 203–212. IEEE (2014)
Acknowledgments
The authors would like to thank CNPq and CAPES for partially funding this work.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
A Appendix
A Appendix
1.1 A.1 Visualization of Geometric Features
Our viewer was developed using Unity3DFootnote 3 engine, with \(C\#\) programming language. The viewer allows users to rewind, accelerate and stop the scene through a time controller and can focus on something interesting multiple times at any time. Figure 1 shows the main window of the viewer. As identified in Fig. 1, the viewer is divided in five parts, as follows: 1) time controller, where is possible to start, stop and continue simulation playback; 2) ChangeScene and RestartCamPos buttons to respectively load data file of another video and restart the camera position for first person view; 3) a window that shows the top view of the environment; 4) first-person view of a previously selected agent (this agent is highlighted in area 3); and 5) which contains the resource panel, where users can enable the visualization of data related to agents’ emotion, socialization and collectivity.
This viewer has three modes of visualization: (i) first-person camera, (ii) top camera, and (iii) an oblique camera. Figure 2 shows an example of each type of camera viewpoint in a video available in Cultural Crowds dataset. In addition to these different viewpoints, it is possible to observe all pedestrians present at each frame f. Pedestrians are represented by a humanoid or cylinder type avatar. Each pedestrian i present in frame f has a position (\(X_i,Y_i\)) (already converted from image coordinates to world coordinates). In addition to positions, it is also possible to know if the pedestrian is walking, running or stopped in frame f through the current speed \(s_i\). If in a certain frame the current speed is greater than or equal to \(\frac{0.08m}{f}\), which is equivalent to \(\frac{2m}{s}\), considering \(\frac{24f}{s}\), then the avatar is running. It was defined based on the Preferred Transition Speed PTS [2] to change from walk to run, for instance. The values of the transitions can be seen in Eq. 1, considering the current speed of the agent \(s_i\).
Also, for the humanoid avatar type, each speed transition is accompanied by an animation transition, for example, if the current speed \(s_i = 0\), then it does not change animation (remaining stationary), but if its speed is \(0< s_i < \frac{0.08m}{f}\), then the animation changes for walking as well as if \(s_i \ge \frac{0.08m}{f}\), the animation of avatar changes to running.
1.2 A.2 Table of Scenes
The Table 1 shows all scenes used in all questions in survey.
1.3 A.3 Questions
The Questionnaire. We formulated a questionnaire to evaluate human perceptions, divided into two stages: i) The first aims to assess whether participants can perceive geometric features in avatars and from different points of view; and ii) the second aims to assess whether participants can perceive emotions and personality traits geometrically, extracted in real videos and represented in avatars. Before the questionnaire, we notified the participants that they could withdraw at any time and for any reason. We do not collect personal data. Before each question, participants saw two to three of the scenes (shown in Table 1) referring to geometric or cultural feature presented in the question. No explanation of the content of the survey was provided to try to avoid bias. In addition, we used Google Forms to apply the questionnaire on social networks, so all participants were volunteers. Next sections described the evaluated aspects.
Geometric Features. The first part of the questionnaire contains twenty-two questions, as shown in Table 2, six related to density, six to speed, five to angular variation and five to distance. Table 2 also presents the possible answers, and the right ones highlighted in bold. In all density questions (Table 2) we asked in which of the short sequences the participant observed the highest density level. The first question (D1) represented a control question, i.e., we wanted to assess whether participants could perceive density variation: low, medium and high density video scenes of pedestrians in crowds. Questions D2 and D3 aimed to assess whether different points of view (camera types) would influence participants’ density perception. Our goal was to assess whether density perception would change due to the camera’s point of view and the way agents were displayed. As shown in Table 2, both questions D2 and D3 scenes were applied after same scene, but displayed with different points of view, and different avatars. Before question D4, two scenes are presented with same density and same point of view, but changing avatar type. Those evaluations aimed to assess whether different types of avatars and point of view would influence density perception. In D5 and D6, walls were added around scenes (4 and 9). D5 had the same objective as D1, i.e., control question. D6 aimed to assess whether walls would influence density perception using first person camera.
Regarding speed perception, all questions had low density. The purpose of these questions was to assess whether participants were able to perceive different speed levels (presented in Eq. 1) in different points of view and types of avatars. The first-person camera was not evaluated for speed questions, as avatars’ first-person views had a lot of variation. The questions \(S1-S4\) aimed to evaluate speed perception over points of view (top and oblique camera). S1 presented two scenes with "run" speed on oblique and top cameras. Same process for S2 but with the "walk" speed. Both S3 and S4 presented two scenes with both speed types. Finally, both S5 and S6 (which aimed to assess the influence of avatar on speed perception) presented two scenes containing two types of avatars using walk speed, one with top camera and one with oblique.
Regarding angular variation and distance, all questions used \(BR-34\) (high density) video scenes. A1, A2, E1 and E2 aimed to assess whether points of view could influence perceptions of angular variation and distance. A1 and E1 presented three scenes with humanoids in three types of cameras. Similar process for A2 and E2, with cylinder type avatars. Finally, A3, A4, A5, E3, E4 and E5 presented, both for angular variation and for distance, two scenes containing two different avatars with top, oblique and first person cameras. These questions aimed to assess whether the types of avatars would influence perceptions.
Cultural Features. The second stage of the questionnaire contains seven questions related to emotions and personality traits. In all scenes presented, two avatars of different colors (red and yellow) were highlighted to be perceptual focus questions, as illustrated in Fig. 3. All questions related to cultural features with their answers (right answers are highlighted in bold) are presented in Table 3. We used as ground truth the results obtained by the approach proposed by Favaretto et al. [13]. In the example of Fig. 3, the initial and final frames of scene15 (shown in Table 1) are shown, where there is a group of pedestrians (represented by avatars) on the right part of the scene. The avatar highlighted in yellow is part of this group and the avatar highlighted in red walk trough the group with a higher speed.
This scene is related to Q1 and Q2, questions that questioned which avatars (yellow or red) were, respectively, neurotic and angry. Q3 and Q4, related to scene10, aimed to assess, respectively, whether the participants would perceive which highlighted avatar was open to new experiences and which one was afraid. Scene10 shows a yellow highlighted avatar interacting with a group of avatars and a red highlighted avatar standing alone with no interaction. Finally, Q5, Q6 and Q7 aimed to assess happiness, extraversion and sociability. Q7 was proposed after analyzing the results of question Q6, so it is explained in Sect. 4.2. These questions are related to scene16, which contains a yellow highlighted avatar walking with a group of avatars and a red highlighted avatar walking alone, in the opposite direction to all other avatars.
1.4 A.4 Table of Results
The Table 4 presents the averages of right answers for each question of our work.
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Araujo, V., Dalmoro, B., Favaretto, R., Vilanova, F., Costa, A., Musse, S.R. (2021). How Much Do We Perceive Geometric Features, Personalities and Emotions in Avatars?. In: Magnenat-Thalmann, N., et al. Advances in Computer Graphics. CGI 2021. Lecture Notes in Computer Science(), vol 13002. Springer, Cham. https://doi.org/10.1007/978-3-030-89029-2_42
Download citation
DOI: https://doi.org/10.1007/978-3-030-89029-2_42
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-89028-5
Online ISBN: 978-3-030-89029-2
eBook Packages: Computer ScienceComputer Science (R0)