skip to main content
research-article

Reproducing Reality: Multimodal Contributions in Natural Scene Discrimination

Published:19 July 2016Publication History
Skip Abstract Section

Abstract

Most research on multisensory processing focuses on impoverished stimuli and simple tasks. In consequence, very little is known about the sensory contributions in the perception of real environments. Here, we presented 23 participants with paired comparison tasks, where natural scenes were discriminated in three perceptually meaningful attributes: movement, openness, and noisiness. The goal was to assess the auditory and visual modality contributions in scene discrimination with short (≤500ms) natural scene exposures. The scenes were reproduced in an immersive audiovisual environment with 3D sound and surrounding visuals. Movement and openness were found to be mainly visual attributes with some input from auditory information. In some scenes, the auditory system was able to derive information about movement and openness that was comparable with audiovisual condition already after 500ms stimulation. Noisiness was mainly auditory, but visual information was found to have a facilitatory role in a few scenes. The sensory weights were highly imbalanced in favor of the stronger modality, but the weaker modality was able to affect the bimodal estimate in some scenes.

References

  1. D. Alais and D. Burr. 2004. The ventriloquist effect results from near-optimal bimodal integration. Current Biology 14, 257--262.Google ScholarGoogle ScholarCross RefCross Ref
  2. Alan Armstrong and Johann Issartel. 2014. Sensorimotor synchronization with audio-visual stimuli: Limited multisensory integration. Experimental Brain Research 232, 11.Google ScholarGoogle ScholarCross RefCross Ref
  3. D. Bates, M. Mächler, B. Bolker, and S. Walker. 2015. Fitting linear mixed-effects models using lme4. Journal of Statistical Software 67, 1, 1--48.Google ScholarGoogle ScholarCross RefCross Ref
  4. Yi Chuan Chen and Charles Spence. 2010. When hearing the bark helps to identify the dog: Semantically-congruent sounds modulate the identification of masked pictures. Cognition 114, 389--404.Google ScholarGoogle ScholarCross RefCross Ref
  5. Yi-Chuan Chen, Su-Ling Yeh, and Charles Spence. 2011. Crossmodal constraints on human perceptual awareness: Auditory semantic modulation of binocular rivalry. Frontiers in Psychology 2, 1--13.Google ScholarGoogle ScholarCross RefCross Ref
  6. Ken Cheng, Sara J. Shettleworth, Janellen Huttenlocher, and John J. Rieser. 2007. Bayesian integration of spatial information. Psychological Bulletin 133, 4, 625--637.Google ScholarGoogle ScholarCross RefCross Ref
  7. Francis B. Colavita. 1974. Human sensory dominance. Perception & Psychophysics 16, 2, 409--412.Google ScholarGoogle ScholarCross RefCross Ref
  8. Verena Conrad, Mario Kleiner, Andreas Bartels, Jessica Hartcher O’Brien, Heinrich H. Bülthoff, and Uta Noppeney. 2013. Naturalistic stimulus structure determines the integration of audiovisual looming signals in binocular rivalry. PLoS One 8, 8, e70710.Google ScholarGoogle ScholarCross RefCross Ref
  9. Beatrice De Gelder and Paul Bertelson. 2003. Multisensory integration, perception and ecological validity. Trends in Cognitive Sciences 7, 10, 460--467.Google ScholarGoogle ScholarCross RefCross Ref
  10. Oliver Doehrmann and Marcus J. Naumer. 2008. Semantics and the multisensory brain: How meaning modulates processes of audio-visual integration. Brain Research 1242, 136--150.Google ScholarGoogle ScholarCross RefCross Ref
  11. A. J. Ecker and L. M. Heller. 2005. Auditory-visual interactions in the perception of a ball’s path. Perception 34, 1, 59--75.Google ScholarGoogle ScholarCross RefCross Ref
  12. Marc O. Ernst and Heinrich H. Bülthoff. 2004. Merging the senses into a robust percept. Trends in Cognitive Sciences 8, 4, 162--169.Google ScholarGoogle ScholarCross RefCross Ref
  13. K. K. Evans and Anne Treisman. 2010. Natural cross-modal mappings between visual and auditory features. Journal of Vision 10, 1, 1--12.Google ScholarGoogle Scholar
  14. J. Gómez Bolaños and V. Pulkki. 2012. Immersive audiovisual environment with 3D audio playback. In 132nd Convention of the Audio Engineering Society. Budapest, Hungary, 1--9.Google ScholarGoogle Scholar
  15. David Hecht and Miriam Reiner. 2009. Sensory dominance in combinations of audio, visual and haptic stimuli. Experimental Brain Research 193, 2, 307--314.Google ScholarGoogle ScholarCross RefCross Ref
  16. A. Koene, Derek Arnold, and Alan Johnston. 2007. Bimodal sensory discrimination is finer than dual single modality discrimination. Journal of Vision 7, 11, 1--11.Google ScholarGoogle ScholarCross RefCross Ref
  17. J. R. Landis and G. G. Koch. 1977. The measurement of observer agreement for categorical data. Biometrics 33, 159--174.Google ScholarGoogle ScholarCross RefCross Ref
  18. Paul J. Laurienti, Robert A. Kraft, Joseph A. Maldjian, Jonathan H. Burdette, and Mark T. Wallace. 2004. Semantic congruence is a critical factor in multisensory behavioral performance. Experimental Brain Research 158, 4, 405--414.Google ScholarGoogle ScholarCross RefCross Ref
  19. J. López-Moliner and S. Soto-Faraco. 2007. Vision affects how fast we hear sounds move. Journal of Vision 7, 12, 1--7.Google ScholarGoogle ScholarCross RefCross Ref
  20. Catarina Mendonça, Olli Rummukainen, and Ville Pulkki. 2015. 3D sound can have a negative impact on the perception of visual content in audiovisual reproductions. In 21st International Conference on Auditory Display (ICAD’15). Graz, Austria.Google ScholarGoogle Scholar
  21. Catarina Mendonça, Jorge A. Santos, and Joan López-Moliner. 2011. The benefit of multisensory integration with biological motion signals. Experimental Brain Research 213, 2--3, 185--192.Google ScholarGoogle ScholarCross RefCross Ref
  22. Alexis Pérez-Bellido, Salvador Soto-Faraco, and Joan López-Moliner. 2013. Sound-driven enhancement of vision: Disentangling detection-level from decision-level contributions. Journal of Neurophysiology 109, 4, 1065--1077.Google ScholarGoogle ScholarCross RefCross Ref
  23. Robert J. Pheasant, Mark N. Fisher, Greg R. Watts, David J. Whitaker, and Kirill V. Horoshenkov. 2010. The importance of auditory-visual interaction in the construction of tranquil space. Journal of Environmental Psychology 30, 4, 501--509.Google ScholarGoogle ScholarCross RefCross Ref
  24. A. Politis and V. Pulkki. 2011. Broadband analysis and synthesis for DirAC using A-format. In 131st Convention of the Audio Engineering Society. New York, NY, 1--11.Google ScholarGoogle Scholar
  25. V. Pulkki. 2007. Spatial sound reproduction with directional audio coding. Journal of the Audio Engineering Society 55, 6, 503--516.Google ScholarGoogle Scholar
  26. R Core Team. 2015. R: A Language and Environment for Statistical Computing. http://www.R-project.org.Google ScholarGoogle Scholar
  27. O. Rummukainen, J. Radun, T. Virtanen, and V. Pulkki. 2014. Categorization of natural dynamic audiovisual scenes. PLoS One 9, 5, e95848.Google ScholarGoogle ScholarCross RefCross Ref
  28. Charles Spence and Sarah Squire. 2003. Multisensory integration: Maintaining the perception of synchrony. Current Biology 13, 519--521.Google ScholarGoogle ScholarCross RefCross Ref
  29. Jye-sheng Tan and Su-ling Yeh. 2015. Audiovisual integration facilitates unconscious visual scene processing. Journal of Experimental Psychology: Human Perception and Performance 41, 5, 1325--1335.Google ScholarGoogle ScholarCross RefCross Ref
  30. Michel Treisman. 1998. Combining information: Probability summation and probability averaging in detection and discrimination. Psychological Methods 3, 2, 252--265.Google ScholarGoogle ScholarCross RefCross Ref
  31. J. Vilkamo, T. Lokki, and V. Pulkki. 2009. Directional audio coding: Virtual microphone-based synthesis and subjective evaluation. Journal of the Audio Engineering Society 57, 9, 709--724.Google ScholarGoogle Scholar
  32. David R. Wozny, Ulrik R. Beierholm, and Ladan Shams. 2008. Human trimodal perception follows optimal statistical inference. Journal of Vision 8, 3, 1--11.Google ScholarGoogle ScholarCross RefCross Ref
  33. Shlomit Yuval-Greenberg and Leon Y. Deouell. 2009. The dog’s meow: Asymmetrical interaction in cross-modal object recognition. Experimental Brain Research 193, 4, 603--614.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Reproducing Reality: Multimodal Contributions in Natural Scene Discrimination

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Transactions on Applied Perception
        ACM Transactions on Applied Perception  Volume 14, Issue 1
        January 2017
        128 pages
        ISSN:1544-3558
        EISSN:1544-3965
        DOI:10.1145/2974018
        Issue’s Table of Contents

        Copyright © 2016 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 19 July 2016
        • Revised: 1 April 2016
        • Accepted: 1 April 2016
        • Received: 1 October 2015
        Published in tap Volume 14, Issue 1

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Research
        • Refereed

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader