Abstract
Following an object’s position relative to oneself is a fundamental functionality required in intelligent real-world interacting robotic systems. This paper presents a computationally efficient vision based 3D tracking system, which can ultimately operate in real-time on autonomous mobile robots in cluttered environments. At the core of the system, two neural inspired event-based dynamic vision sensors (eDVS) independently track a high frequency flickering LED in their respective 2D angular coordinate frame. A self-adjusted feed-forward neural network maps those independent 2D angular coordinates into a Cartesian 3D position in world coordinates. During an initial calibration phase, an object composed of multiple independent markers with known geometry provides relative position information between those markers for network training (without ever using absolute world coordinates for training). In a subsequent application phase tracking a single marker yields position estimates relative to sensor origin, while tracking multiple markers provides additional orientation. The neural inspired vision-based tracking system runs in real-time on ARM7 microcontrollers, without the need for an external PC.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Andriluka, M., Roth, S., Schiele, B.: Monocular 3d pose estimation and tracking by detection. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA (2010)
Cyr, C.M., Kimia, B.B.: 3D object recognition using shape similarity-based aspect graph. In: Proc. of Eighth IEEE International Conference on Computer Vision (ICCV), Vancouver, BC, Canada (2001)
Ponce, J., et al.: Toward true 3D object recognition. In: Congrès de Reconnaissance des Formes et Intelligence Artificielle, Toulouse, France (2004)
Chao, C.H., et al.: Real-time target tracking and obstacle avoidance for mobile robots using two cameras. In: ICCAS-SICE, pp. 4347–4352. IEEE, Fukuoka (2009)
Shibata, M., Kobayashi, N.: Image-based visual tracking for moving targets with active stereo vision robot. In: International Joint Conference SICE-ICASE, Busan (2006)
Lichtsteiner, P., Posch, C., Delbruck, T.: A 128×128 120dB 15us Latency Asynchronous Temporal Contrast Vision Sensor. IEEE Journal of Solid State Circuits 43(2), 566–576 (2007)
Institute of Neuroinformatics, Silicon Retina Information, http://siliconretina.ini.uzh.ch
Conradt, J., Berner, R., Cook, M., Delbruck, T.: Ano Embedded AER Dynamic Vision Sensor for Low-Latency Ple Balancing. In: IEEE Workshop on Embedded Computer Vision, Kyoto, Japan (2009)
Müller, G.R., Conradt, J.: A Miniature Low-Power Sensor System for Real Time 2D Visual Tracking of LED Markers. In: Proc. IEEE Intl. Conference on Robotics and Biomimetics (ROBIO), Phuket (2011)
Arbib, M.: Handbook of Brain Theory and Neural Networks. Bradford Book (2002) ISBN 0262011972
NaturalPoint, Inc., http://www.naturalpoint.com/optitrack
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Müller, G.R., Conradt, J. (2012). Self-calibrating Marker Tracking in 3D with Event-Based Vision Sensors. In: Villa, A.E.P., Duch, W., Érdi, P., Masulli, F., Palm, G. (eds) Artificial Neural Networks and Machine Learning – ICANN 2012. ICANN 2012. Lecture Notes in Computer Science, vol 7552. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-33269-2_40
Download citation
DOI: https://doi.org/10.1007/978-3-642-33269-2_40
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-33268-5
Online ISBN: 978-3-642-33269-2
eBook Packages: Computer ScienceComputer Science (R0)