Skip to main content

Robust Object Recognition Under Partial Occlusions Using an RGB-D Camera

  • Conference paper
Book cover Robot Intelligence Technology and Applications 3

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 345))

  • 3838 Accesses

Abstract

For a robot to execute a specific task, the robot firstly has to understand what objects are in robot’s view. To complete a specific task in a given time, the computation time for recognition is also important. There are much research for increasing recognition accuracy, but the recognition speed is not enough to be applied in real environment. On the other hand, there are also much research for reducing the computation time for recognition, but the recognition accuracy needs to be further improved. Nowadays, deep network has come into the spotlight due to its speed and accuracy. Deep network doesn’t need to find hand-tuned features. This paper proposes a deep network-based object recognition algorithm. The main contribution is that objects could be recognized under occlusion, as objects are often laid to overlap each other. The occlusion makes object recognition accuracy worse. To overcome this problem, the dataset for training consists of not full images but partial information of images and corresponding ground truths. The object region could be found very quickly by using an RGB-D camera. By assuming that most objects are on the stable plane, object regions are taken easily. Experimental results demonstrate such consideration of contextual information (e.g. objects are on the table) makes the performance of recognition better.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Lai, K., et al.: A large-scale hierarchical multi-view rgb-d object dataset. In: 2011 IEEE International Conference on Robotics and Automation (ICRA). IEEE (2011)

    Google Scholar 

  2. Huang, A.S., et al.: Visual odometry and mapping for autonomous flight using an RGB-D camera. In: International Symposium on Robotics Research, ISRR (2011)

    Google Scholar 

  3. Kim, D.-H., Kim, J.-H.: Image-Based ICP Algorithm for Visual Odometry Using a RGB-D Sensor in a Dynamic Environment. In: Kim, J.-H., Matson, E., Myung, H., Xu, P. (eds.) Robot Intelligence Technology and Applications. AISC, vol. 208, pp. 423–430. Springer, Heidelberg (2013)

    Google Scholar 

  4. Henry, P., et al.: RGB-D mapping: Using depth cameras for dense 3D modeling of indoor environments. In: The 12th International Symposium on Experimental Robotics, ISER (2010)

    Google Scholar 

  5. Lenz, I., Lee, H., Saxena, A.: Deep learning for detecting robotic grasps. arXiv preprint arXiv:1301.3592 (2013)

    Google Scholar 

  6. Helmer, S., Lowe, D.: Using stereo for object recognition. In: 2010 IEEE International Conference on Robotics and Automation (ICRA). IEEE (2010)

    Google Scholar 

  7. Bengio, Y.: Learning deep architectures for AI. Foundations and Trends® in Machine Learning 2(1), 1–127 (2009)

    Article  MATH  MathSciNet  Google Scholar 

  8. Hinton, G.E.: Deep belief networks. Scholarpedia 4(5), 5947 (2009)

    Article  Google Scholar 

  9. Lowe, D.G.: Object recognition from local scale-invariant features. In: The Proceedings of the Seventh IEEE International Conference on Computer Vision, vol. 2. IEEE (1999)

    Google Scholar 

  10. Bay, H., Tuytelaars, T., Van Gool, L.: SURF: Speeded up robust features. In: Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006, Part I. LNCS, vol. 3951, pp. 404–417. Springer, Heidelberg (2006)

    Chapter  Google Scholar 

  11. Fischler, M.A., Bolles, R.C.: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Communications of the ACM 24(6), 381–395 (1981)

    Article  MathSciNet  Google Scholar 

  12. Salakhutdinov, R., Hinton, G.E.: Deep boltzmann machines. In: International Conference on Artificial Intelligence and Statistics (2009)

    Google Scholar 

  13. Bengio, Y., Delalleau, O.: Justifying and generalizing contrastive divergence. Neural Computation 21(6), 1601–1621 (2009)

    Article  MATH  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yong-Ho Yoo .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this paper

Cite this paper

Yoo, YH., Kim, JH. (2015). Robust Object Recognition Under Partial Occlusions Using an RGB-D Camera. In: Kim, JH., Yang, W., Jo, J., Sincak, P., Myung, H. (eds) Robot Intelligence Technology and Applications 3. Advances in Intelligent Systems and Computing, vol 345. Springer, Cham. https://doi.org/10.1007/978-3-319-16841-8_58

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-16841-8_58

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-16840-1

  • Online ISBN: 978-3-319-16841-8

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics