Skip to main content

Where Do We Grasp Objects? – An Experimental Verification of the Selective Attention for Action Model (SAAM)

  • Conference paper
Attention in Cognitive Systems (WAPCV 2008)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 5395))

Included in the following conference series:

Abstract

Classically, visual attention is assumed to be influenced by visual properties of objects, e.g. as assessed in visual search tasks. However, recent experimental evidence suggests that visual attention is also guided by action-related properties of objects (“affordances”,[1,2]), e.g. the handle of a cup affords grasping the cup; therefore attention is drawn towards the handle (see [3] for example). In a first step towards modelling this interaction between attention and action, we implemented the Selective Attention for Action model (SAAM). The design of SAAM is based on the Selective Attention for Identification model (SAIM [4]). For instance, we also followed a soft-constraint satisfaction approach in a connectionist framework. However, SAAM’s selection process is guided by locations within objects suitable for grasping them whereas SAIM selects objects based on their visual properties. In order to implement SAAM’s selection mechanism two sets of constraints were implemented. The first set of constraints took into account the anatomy of the hand, e.g. maximal possible distances between fingers. The second set of constraints (geometrical constraints) considered suitable contact points on objects by using simple edge detectors. At first, we demonstrate here that SAAM can successfully mimic human behaviour by comparing simulated contact points with experimental data. Secondly, we show that SAAM simulates affordance-guided attentional behaviour as it successfully generates contact points for only one object in two-object images.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Gibson, J.J.: The senses considered as perceptual systems. Houghton-Mifflin, Boston (1966)

    Google Scholar 

  2. Gibson, J.J.: The ecological approach to visual perception. Houghton-Mifflin, Boston (1979)

    Google Scholar 

  3. di Pellegrino, G., Rafal, R., Tipper, S.P.: Implicitly evoked actions modulate visual selection: evidence from parietal extinction. Current Biology 15(16), 1469–1472 (2005)

    Article  PubMed  Google Scholar 

  4. Heinke, D., Humphreys, G.W.: Attention, spatial representation and visual neglect: Simulating emergent attention and spatial memory in the selective attention for identication model (SAIM). Psychological Review 110(1), 29–87 (2003)

    Article  PubMed  Google Scholar 

  5. Grafton, S.T., Fadiga, L., Arbib, M.A., Rizzolatti, G.: Premotor cortex activation during observation and naming of familiar tools. NeuroImage 6(4), 231–236 (1997)

    Article  CAS  PubMed  Google Scholar 

  6. Grèzes, J., Decety, J.: Does visual perception of objects afford action? evidence from a neuroimaging study. Neuropsychologia 40(2), 212–222 (2002)

    Article  PubMed  Google Scholar 

  7. Tucker, M., Ellis, R.: On the relations between seen objects and components of potential actions. Journal of Experimental Psychology 24(3), 830–846 (1998)

    CAS  PubMed  Google Scholar 

  8. Phillips, J.C., Ward, R.: S–r correspondence effects of irrelevant visual affordance: Time course and specificity of response activation. Visual Cognition 9(4–5), 540–558 (2002)

    Article  Google Scholar 

  9. Borghi, A.M., Bonfiglioli, C., Lugli, L., Ricciardelli, P., Rubichi, S., Nicoletti, R.: Are visual stimuli sufficient to evoke motor information? studies with hand primes. Neuroscience Letters 411(1), 17–21 (2007)

    Article  CAS  PubMed  Google Scholar 

  10. Handy, T.C., Grafton, S.T., Shroff, N.M., Ketay, S., Gazzaniga, M.S.: Graspable objects grab attention when the potential for action is recognized. Nature Neuroscience 6(4), 421–427 (2003)

    Article  CAS  PubMed  Google Scholar 

  11. Hopfield, J.J., Tank, D.W.: “neural” computation of decisions in optimization problems. Biological Cybernetics 52(3), 141–152 (1985)

    CAS  PubMed  Google Scholar 

  12. Gonzalez, R.C., Woods, R.E.: Digital Image Processing. Addison-Wesley, Reading (1993)

    Google Scholar 

  13. Hopfield, J.J.: Neural networks and physical systems with emergent collective computational abilities. Proceedings of the National Academy of Sciences, vol. 79, pp. 2554–2558 (1982)

    Google Scholar 

  14. Mjolsness, E., Garrett, C.: Algebraic transformations of objective functions. Neural Networks 3(6), 651–669 (1990)

    Article  Google Scholar 

  15. Hindmarsh, A.C., Brown, P.N., Grant, K.E., Lee, S.N., Serban, R., Shumaker, D.E., Woodward, C.S.: Sundials: Suite of nonlinear and differential/algebraic equation solvers. ACM Transactions on Mathematical Software 31(3), 363–396 (2005); also available as LLNL technical report UCRL-JP-200037

    Article  Google Scholar 

  16. Carey, D.P., Harvey, M., Milner, A.D.: Visuomotor sensitivity for shape and orientation in a patient with visual form agnosia. Neuropsychologia 34(5), 329–337 (1996)

    Article  CAS  PubMed  Google Scholar 

  17. Mason, M.T.: Mechanics of Robotic Manipulation. Intelligent Robots and Autonomous Agents. MIT Press, Cambridge (2001)

    Google Scholar 

  18. Humphreys, G.W., Riddoch, M.J.: From vision to action and action to vision: a convergent route approach to vision, action, and attention. Psychology of learning and motivation 42, 225–264 (2003)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2009 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Böhme, C., Heinke, D. (2009). Where Do We Grasp Objects? – An Experimental Verification of the Selective Attention for Action Model (SAAM). In: Paletta, L., Tsotsos, J.K. (eds) Attention in Cognitive Systems. WAPCV 2008. Lecture Notes in Computer Science(), vol 5395. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-00582-4_4

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-00582-4_4

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-00581-7

  • Online ISBN: 978-3-642-00582-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics