skip to main content
10.1145/3387168.3387235acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicvispConference Proceedingsconference-collections
research-article

Curiosity-Driven Exploration Effectiveness on Various Environments

Authors Info & Claims
Published:25 May 2020Publication History

ABSTRACT

Hand Crafting Reward functions have never been scalable solutions for real world problems. The self-generated intrinsic rewards inspired by human curiosity may be one of scalable answers to solve sparse reward problem. The research thus investigated the effectiveness of some selected techniques based on the theory of curiosity-driven exploration. The Count-based, Prediction-based and other methods in total of six algorithms were experimented on various OpenAI gym environments. The results showed that the exploration algorithms have an impact on software agent in ability to find optimal solutions compared with the baseline in many cases. Still, there is no clear winner between the selected exploration methods and the best scalable exploration is not yet explored. The finding is that the added small intrinsic reward noise helps improve sample efficiency in the short run.

References

  1. A. Aubret, L. Matignon, and S. Hassas, "A survey on intrinsic motivation in reinforcement learning," arXiv preprint arXiv:1908.06976, 2019.Google ScholarGoogle Scholar
  2. B. Baker et al., "Emergent tool use from multi-agent autocurricula," arXiv preprint arXiv:1909.07528, 2019.Google ScholarGoogle Scholar
  3. G. Brockman et al., "Openai gym," arXiv preprint arXiv:1606.01540, 2016.Google ScholarGoogle Scholar
  4. Y. Burda, H. Edwards, D. Pathak, A. Storkey, T. Darrell, and A. A. Efros, "Large-scale study of curiosity-driven learning," arXiv preprint arXiv:1808.04355, 2018.Google ScholarGoogle Scholar
  5. Y. Burda, H. Edwards, A. Storkey, and O. Klimov, "Exploration by random network distillation," arXiv preprint arXiv:1810.12894, 2018.Google ScholarGoogle Scholar
  6. M. Cholodovskis Machado, "Efficient Exploration in Reinforcement Learning through Time-Based Representations," 2019.Google ScholarGoogle Scholar
  7. T. G. Dietterich, "The MAXQ Method for Hierarchical Reinforcement Learning.," in ICML, 1998, vol. 98, pp. 118--126.Google ScholarGoogle Scholar
  8. R. Houthooft, X. Chen, Y. Duan, J. Schulman, F. De Turck, and P. Abbeel, "Vime: Variational information maximizing exploration," in Advances in Neural Information Processing Systems, 2016, pp. 1109--1117. [2] F.N.M Surname, Article Title, https://www.acm.org/publications/proceedings-template.Google ScholarGoogle Scholar
  9. S. Iqbal and F. Sha, "Coordinated Exploration via Intrinsic Rewards for Multi-Agent Reinforcement Learning," arXiv preprint arXiv:1905.12127, 2019.Google ScholarGoogle Scholar
  10. M. Jaderberg et al., "Reinforcement learning with unsupervised auxiliary tasks," arXiv preprint arXiv:1611.05397, 2016.Google ScholarGoogle Scholar
  11. OpenAI, "Gym: A toolkit for developing and comparing reinforcement learning algorithms." [Online]. Available: https://gym.openai.com. [Accessed: 13-Oct-2019].Google ScholarGoogle Scholar
  12. D. Pathak, P. Agrawal, A. A. Efros, and T. Darrell, "Curiosity-driven exploration by self-supervised prediction," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2017, pp. 16--17.Google ScholarGoogle Scholar
  13. H. Tang et al., "# Exploration: A study of count-based exploration for deep reinforcement learning," in Advances in neural information processing systems, 2017, pp. 2753--2762.Google ScholarGoogle Scholar
  14. S. B. Thrun, "Efficient exploration in reinforcement learning," 1992.Google ScholarGoogle Scholar

Index Terms

  1. Curiosity-Driven Exploration Effectiveness on Various Environments

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      ICVISP 2019: Proceedings of the 3rd International Conference on Vision, Image and Signal Processing
      August 2019
      584 pages
      ISBN:9781450376259
      DOI:10.1145/3387168

      Copyright © 2019 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 25 May 2020

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited

      Acceptance Rates

      ICVISP 2019 Paper Acceptance Rate126of277submissions,45%Overall Acceptance Rate186of424submissions,44%
    • Article Metrics

      • Downloads (Last 12 months)5
      • Downloads (Last 6 weeks)0

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader