skip to main content
10.1145/3449726.3459421acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
poster

Growth and harvest induce essential dynamics in neural networks

Published:08 July 2021Publication History

ABSTRACT

Training neural networks with faster gradient methods brings them to the edge of stability, proximity to which improves their generalization capability. However, it is not clear how to stably approach the edge. We propose a new activation function to model inner processes inside neurons with single-species population dynamics. The function induces essential dynamics in neural networks with a growth and harvest rate to improve their generalization capability.

References

  1. Sanjeev Arora, Nadav Cohen, Noah Golowich, and Wei Hu. 2019. A convergence analysis of gradient descent for deep linear neural networks. In ICLR.Google ScholarGoogle Scholar
  2. Garrett Bingham, William Macke, and Risto Miikkulainen. 2020. Evolutionary optimization of deep learning activation functions. In Proceedings of the 2020 Genetic and Evolutionary Computation Conference (GECCO '20). 289--296.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Valentin De Bortoli, Alain Durmus, Xavier Fontaine, and Umut Simsekli. 2020. Quantitative Propagation of Chaos for SGD in Wide Neural Networks. In NeurIPS.Google ScholarGoogle Scholar
  4. Jeremy Cohen, Simran Kaur, Yuanzhi Li, J Zico Kolter, and Ameet Talwalkar. 2021. Gradient Descent on Neural Networks Typically Occurs at the Edge of Stability. In ICLR.Google ScholarGoogle Scholar
  5. Aymeric Dieuleveut, Nicolas Flammarion, and Francis Bach. 2017. Harder, Better, Faster, Stronger Convergence Rates for Least-Squares Regression. Journal of Machine Learning Research 18 (2017), 1--51.Google ScholarGoogle Scholar
  6. P. De Felice, C. Marangi, G. Nardulli, G. Pasquariello, and L. Tedesco. 1993. Dynamics of neural networks with non-monotone activation function. Network: Computation in Neural Systems 4, 1 (1993), 1--9.Google ScholarGoogle ScholarCross RefCross Ref
  7. William G. Gray and Genetha A. Gray (Eds.). 2017. Introduction to Environmental Modeling. Cambridge University Press, Cambridge, UK.Google ScholarGoogle Scholar
  8. T. Legović. 2016. Dynamic population models. In Ecological model types, S.E. Jorgensen (Ed.). Elsevier, 39--63.Google ScholarGoogle Scholar
  9. Zhiyuan Li and Sanjeev Arora. 2020. An Exponential Learning Rate Schedule for Deep Learning. In ICLR.Google ScholarGoogle Scholar
  10. Samuel S. Schoenholz, Justin Gilmer, Surya Ganguli, and Jascha Sohl-Dickstein. 2017. Deep information propagation. In ICLR.Google ScholarGoogle Scholar
  11. Andrew Trask, Felix Hill, Scott Reed, Jack Rae, Chris Dyer, and Phil Blunsom. 2018. Neural Arithmetic Logic Units. In NeurIPS.Google ScholarGoogle Scholar
  12. Kees van den Doel and Uri Ascher. 2012. The Chaotic Nature of Faster Gradient Descent Methods. Journal of Scientific Computing 51 (2012), 560--581.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Rui Wang, Danielle Robinson, Christos Faloutsos, Yuyang Wang, and Rose Yu. 2020. Learning dynamical systems requires rethinking generalization. In 1st NeurIPS workshop on Interpretable Inductive Biases and Physically Structured Learning.Google ScholarGoogle Scholar

Index Terms

  1. Growth and harvest induce essential dynamics in neural networks

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      GECCO '21: Proceedings of the Genetic and Evolutionary Computation Conference Companion
      July 2021
      2047 pages
      ISBN:9781450383516
      DOI:10.1145/3449726

      Copyright © 2021 Owner/Author

      Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 8 July 2021

      Check for updates

      Qualifiers

      • poster

      Acceptance Rates

      Overall Acceptance Rate1,669of4,410submissions,38%

      Upcoming Conference

      GECCO '24
      Genetic and Evolutionary Computation Conference
      July 14 - 18, 2024
      Melbourne , VIC , Australia
    • Article Metrics

      • Downloads (Last 12 months)1
      • Downloads (Last 6 weeks)0

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader