Skip to main content

Adaptive Stochastic Natural Gradient Method for Optimizing Functions with Low Effective Dimensionality

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 12269))

Abstract

Black-box optimization algorithms, such as evolutionary algorithms, have been recognized as useful tools for real-world applications. Several efficient probabilistic model-based evolutionary algorithms, such as the compact genetic algorithm (cGA) and the covariance matrix adaptation evolution strategy (CMA-ES), can be regarded as a stochastic natural gradient ascent on statistical manifolds. Our baseline algorithm is the adaptive stochastic natural gradient (ASNG) method which automatically adapts the learning rate based on the signal-to-noise ratio (SNR) of the approximated natural gradient. ASNG has shown effectiveness in a practical application, but the convergence speed of ASNG deteriorates on objective functions with low effective dimensionality (LED), where LED means that part of the design variables is ineffective or does not affect the objective value significantly. In this paper, we propose an element-wise adjustment method for the approximated natural gradient based on the element-wise SNR and introduce the proposed adjustment method into ASNG. The proposed method suppresses the natural gradient elements with the low SNRs, helping to accelerate the learning rate adaptation in ASNG. We incorporate the proposed method into the cGA and demonstrate the effectiveness of the proposed method on the benchmark functions of binary optimization.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Akimoto, Y., Shirakawa, S., Yoshinari, N., Uchida, K., Saito, S., Nishida, K.: Adaptive stochastic natural gradient method for one-shot neural architecture search. In: Proceedings of the 36th International Conference on Machine Learning (ICML), pp. 171–180 (2019)

    Google Scholar 

  2. Amari, S.: Natural gradient works efficiently in learning. Neural Comput. 10, 251–276 (1998)

    Article  Google Scholar 

  3. Baluja, S., Caruana, R.: Removing the genetics from the standard genetic algorithm. In: Proceedings of the 12th International Conference on Machine Learning (ICML), pp. 38–46 (1995)

    Google Scholar 

  4. Bergstra, J., Bengio, Y.: Random search for hyper-parameter optimization. J. Mach. Learn. Res. 13, 281–305 (2012)

    MathSciNet  MATH  Google Scholar 

  5. Hansen, N., Müller, S.D., Koumoutsakos, P.: Reducing the time complexity of the derandomized evolution strategy with covariance matrix adaptation (CMA-ES). Evol. Comput. 11(1), 1–18 (2003)

    Article  Google Scholar 

  6. Harik, G.R., Lobo, F.G., Goldberg, D.E.: The compact genetic algorithm. IEEE Trans. Evol. Comput. 3, 287–297 (1999)

    Article  Google Scholar 

  7. Hutter, F., Hoos, H., Leyton-Brown, K.: An efficient approach for assessing hyperparameter importance. In: Proceedings of the 31st International Conference on Machine Learning (ICML), pp. 754–762 (2014)

    Google Scholar 

  8. Lukaczyk, T., Constantine, P., Palacios, F., Alonso, J.: Active subspaces for shape optimization. In: Proceedings of the 10th AIAA Multidisciplinary Design Optimization Conference, pp. 1–18 (2014)

    Google Scholar 

  9. Ollivier, Y., Arnold, L., Auger, A., Hansen, N.: Information-geometric optimization algorithms: a unifying picture via invariance principles. J. Mach. Learn. Res. 18, 1–65 (2017)

    MathSciNet  MATH  Google Scholar 

  10. Sanyang, M.L., Kabán, A.: REMEDA: random embedding EDA for optimising functions with intrinsic dimension. In: Handl, J., Hart, E., Lewis, P.R., López-Ibáñez, M., Ochoa, G., Paechter, B. (eds.) PPSN 2016. LNCS, vol. 9921, pp. 859–868. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-45823-6_80

    Chapter  Google Scholar 

  11. Wang, Z., Hutter, F., Zoghi, M., Matheson, D., de Freitas, N.: Bayesian optimization in a billion dimensions via random embeddings. J. Artif. Intell. Res. 55, 361–387 (2016)

    Article  MathSciNet  Google Scholar 

  12. Wierstra, D., Schaul, T., Glasmachers, T., Sun, Y., Peters, J., Schmidhuber, J.: Natural evolution strategies. J. Mach. Learn. Res. 15, 949–980 (2014)

    MathSciNet  MATH  Google Scholar 

Download references

Acknowledgments

This work is partially supported by the SECOM Science and Technology Foundation and JSPS KAKENHI Grant Number JP20H04240.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Teppei Yamaguchi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yamaguchi, T., Uchida, K., Shirakawa, S. (2020). Adaptive Stochastic Natural Gradient Method for Optimizing Functions with Low Effective Dimensionality. In: Bäck, T., et al. Parallel Problem Solving from Nature – PPSN XVI. PPSN 2020. Lecture Notes in Computer Science(), vol 12269. Springer, Cham. https://doi.org/10.1007/978-3-030-58112-1_50

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-58112-1_50

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-58111-4

  • Online ISBN: 978-3-030-58112-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics