Skip to main content

Viewpoint Planning Based on Uncertainty Maps Created from the Generative Query Network

  • Conference paper
  • First Online:
Book cover Advances in Artificial Intelligence (JSAI 2020)

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 1357))

Included in the following conference series:

  • 323 Accesses

Abstract

Most current research on robotic tasks such as grasping uses the single fixed viewpoint positioned to oversee the environment from above. However, there are cases when a single fixed viewpoint does not provide sufficient information to perform a designated robot task. Environments with high occlusion, or cases where objects have fragile components, it is difficult for robots to identify the task relevant viewpoint to achieve a task or goal. Herein, to reconstruct the world model with less effort, we propose a viewpoint planner method that uses uncertainty in scene representation from Generative Query Network(GQN) With this scene representation, we create an uncertainty map by calculating the pixel-wise variance of multiple predicted images for each query viewpoint. The results indicate that our solution is capable of taking a suitable view of an untrained object, therefore reducing the uncertainty. A suitable viewpoint is defined as one that improves prediction certainty by focusing an area in which the learning environment allows the highest uncertainty value. In a future study, we would like to propose a task specific viewpoint planner models for robotic tasks such as grasping objects with occlusion by extending this research.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 149.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 199.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://github.com/deepmind/gqn-datasets.

References

  1. Bajcsy, R.: Active perception. Proc. IEEE 76(8), 966–1005 (1988)

    Article  Google Scholar 

  2. Bohg, J., Morales, A., Asfour, T., Kragic, D.: Data-driven grasp synthesis-a survey. IEEE Trans. Rob. 30(2), 289–309 (2014)

    Article  Google Scholar 

  3. Chen, S., Li, Y., Kwok, N.: Active vision in robotic systems: a survey of recent developments. Int. J. Robot. Res. 30, 1343–1377 (2011)

    Article  Google Scholar 

  4. Eslami, S.M.A., et al.: Neural scene representation and rendering. Science 360(6394), 1204–1210 (2018)

    Article  Google Scholar 

  5. Gregor, K., Danihelka, I., Graves, A., Wierstra, D.: DRAW: a recurrent neural network for image generation. CoRR, abs/1502.04623 (2015)

    Google Scholar 

  6. Kappler, D., Bohg, J., Schaal, S.: Leveraging big data for grasp planning. In: 2015 IEEE International Conference on Robotics and Automation (ICRA), pp. 4304–4311 (2015)

    Google Scholar 

  7. Lenz, I., Lee, H., Saxena, A.: Deep learning for detecting robotic grasps. Int. J. Robot. Res. 34(4–5), 705–724 (2015)

    Article  Google Scholar 

  8. Levine, S., Pastor, P., Krizhevsky, A., Ibarz, J., Quillen, D.: Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection. Int. J. Robot. Res. 37(4–5), 421–436 (2018)

    Article  Google Scholar 

  9. Mahler, J., et al.: Dex-Net 2.0: deep learning to plan robust grasps with synthetic point clouds and analytic grasp metrics (2017)

    Google Scholar 

  10. Mahler, J., Matl, M., Liu, X., Li, A., Gealy, D., Goldberg, K.: Dex-Net 3.0: computing robust robot suction grasp targets in point clouds using a new analytic model and deep learning. arXiv preprint arXiv:1709.06670 (2017)

  11. Mahler, J., et al.: Learning ambidextrous robot grasping policies. Sci. Robot. 4(26), eaau4984 (2019)

    Google Scholar 

  12. Manuelli, L., Gao, W., Florence, P., Tedrake, R.: kPAM: keypoint affordances for category-level robotic manipulation. arXiv, abs/1903.06684 (2019

    Google Scholar 

  13. Nguyen-Phuoc, T., Li, C., Theis, L., Richardt, C., Yang, Y.: HoloGAN: unsupervised learning of 3D representations from natural images. CoRR, abs/1904.01326 (2019)

    Google Scholar 

  14. Qiu, Y., Satoh, Y., Suzuki, R., Iwata, K., Kataoka, H.: Multi-view visual question answering with active viewpoint selection. Sensors 20(8), 2281 (2020)

    Article  Google Scholar 

  15. Rasouli, A., Tsotsos, J.K.: The effect of color space selection on detectability and discriminability of colored objects. CoRR, abs/1702.05421 (2017)

    Google Scholar 

  16. Sather, J., Zhang, X.J.: Viewpoint optimization for autonomous strawberry harvesting with deep reinforcement learning. arXiv, abs/1903.02074 (2019)

    Google Scholar 

  17. Sitzmann, V., Zollhöfer, M., Wetzstein, G.: Scene representation networks: continuous 3D-structure-aware neural scene representations. CoRR, abs/1906.01618 (2019)

    Google Scholar 

  18. Tobin, J., OpenAI, Abbeel, P.: Geometry-aware neural rendering. arXiv, abs/1911.04554 (2019)

    Google Scholar 

Download references

Acknowledgment

This work was supported by Japan Science and Technology Agency (JST) CREST Grant Number JPMJCR15E3 “Symbol Emergence in Robotics for Future Human-Machine Collaboration”, Japan.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kelvin Lukman .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Lukman, K., Mori, H., Ogata, T. (2021). Viewpoint Planning Based on Uncertainty Maps Created from the Generative Query Network. In: Yada, K., et al. Advances in Artificial Intelligence. JSAI 2020. Advances in Intelligent Systems and Computing, vol 1357. Springer, Cham. https://doi.org/10.1007/978-3-030-73113-7_4

Download citation

Publish with us

Policies and ethics