Abstract
For a robot to execute a specific task, the robot firstly has to understand what objects are in robot’s view. To complete a specific task in a given time, the computation time for recognition is also important. There are much research for increasing recognition accuracy, but the recognition speed is not enough to be applied in real environment. On the other hand, there are also much research for reducing the computation time for recognition, but the recognition accuracy needs to be further improved. Nowadays, deep network has come into the spotlight due to its speed and accuracy. Deep network doesn’t need to find hand-tuned features. This paper proposes a deep network-based object recognition algorithm. The main contribution is that objects could be recognized under occlusion, as objects are often laid to overlap each other. The occlusion makes object recognition accuracy worse. To overcome this problem, the dataset for training consists of not full images but partial information of images and corresponding ground truths. The object region could be found very quickly by using an RGB-D camera. By assuming that most objects are on the stable plane, object regions are taken easily. Experimental results demonstrate such consideration of contextual information (e.g. objects are on the table) makes the performance of recognition better.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Lai, K., et al.: A large-scale hierarchical multi-view rgb-d object dataset. In: 2011 IEEE International Conference on Robotics and Automation (ICRA). IEEE (2011)
Huang, A.S., et al.: Visual odometry and mapping for autonomous flight using an RGB-D camera. In: International Symposium on Robotics Research, ISRR (2011)
Kim, D.-H., Kim, J.-H.: Image-Based ICP Algorithm for Visual Odometry Using a RGB-D Sensor in a Dynamic Environment. In: Kim, J.-H., Matson, E., Myung, H., Xu, P. (eds.) Robot Intelligence Technology and Applications. AISC, vol. 208, pp. 423–430. Springer, Heidelberg (2013)
Henry, P., et al.: RGB-D mapping: Using depth cameras for dense 3D modeling of indoor environments. In: The 12th International Symposium on Experimental Robotics, ISER (2010)
Lenz, I., Lee, H., Saxena, A.: Deep learning for detecting robotic grasps. arXiv preprint arXiv:1301.3592 (2013)
Helmer, S., Lowe, D.: Using stereo for object recognition. In: 2010 IEEE International Conference on Robotics and Automation (ICRA). IEEE (2010)
Bengio, Y.: Learning deep architectures for AI. Foundations and Trends® in Machine Learning 2(1), 1–127 (2009)
Hinton, G.E.: Deep belief networks. Scholarpedia 4(5), 5947 (2009)
Lowe, D.G.: Object recognition from local scale-invariant features. In: The Proceedings of the Seventh IEEE International Conference on Computer Vision, vol. 2. IEEE (1999)
Bay, H., Tuytelaars, T., Van Gool, L.: SURF: Speeded up robust features. In: Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006, Part I. LNCS, vol. 3951, pp. 404–417. Springer, Heidelberg (2006)
Fischler, M.A., Bolles, R.C.: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Communications of the ACM 24(6), 381–395 (1981)
Salakhutdinov, R., Hinton, G.E.: Deep boltzmann machines. In: International Conference on Artificial Intelligence and Statistics (2009)
Bengio, Y., Delalleau, O.: Justifying and generalizing contrastive divergence. Neural Computation 21(6), 1601–1621 (2009)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Yoo, YH., Kim, JH. (2015). Robust Object Recognition Under Partial Occlusions Using an RGB-D Camera. In: Kim, JH., Yang, W., Jo, J., Sincak, P., Myung, H. (eds) Robot Intelligence Technology and Applications 3. Advances in Intelligent Systems and Computing, vol 345. Springer, Cham. https://doi.org/10.1007/978-3-319-16841-8_58
Download citation
DOI: https://doi.org/10.1007/978-3-319-16841-8_58
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-16840-1
Online ISBN: 978-3-319-16841-8
eBook Packages: EngineeringEngineering (R0)