Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3321707.3321876acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
research-article

Evolvability ES: scalable and direct optimization of evolvability

Published: 13 July 2019 Publication History
  • Get Citation Alerts
  • Abstract

    Designing evolutionary algorithms capable of uncovering highly evolvable representations is an open challenge in evolutionary computation; such evolvability is important in practice, because it accelerates evolution and enables fast adaptation to changing circumstances. This paper introduces evolvability ES, an evolutionary algorithm designed to explicitly and efficiently optimize for evolvability, i.e. the ability to further adapt. The insight is that it is possible to derive a novel objective in the spirit of natural evolution strategies that maximizes the diversity of behaviors exhibited when an individual is subject to random mutations, and that efficiently scales with computation. Experiments in 2-D and 3-D locomotion tasks highlight the potential of evolvability ES to generate solutions with tens of thousands of parameters that can quickly be adapted to solve different tasks and that can productively seed further evolution. We further highlight a connection between evolvability in EC and a recent and popular gradient-based meta-learning algorithm called MAML; results show that evolvability ES can perform competitively with MAML and that it discovers solutions with distinct properties. The conclusion is that evolvability ES opens up novel research directions for studying and exploiting the potential of evolvable representations for deep neural networks.

    Supplementary Material

    PDF File (p107-gajewski_suppl.pdf)
    Supplemental material.

    References

    [1]
    Martín Abadi, Ashish Agarwal, Paul Barham, Eugene Brevdo, Zhifeng Chen, Craig Citro, Greg S. Corrado, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Ian Goodfellow, Andrew Harp, Geoffrey Irving, Michael Isard, Yangqing Jia, Rafal Jozefowicz, Lukasz Kaiser, Manjunath Kudlur, Josh Levenberg, Dan Mané, Rajat Monga, Sherry Moore, Derek Murray, Chris Olah, Mike Schuster, Jonathon Shlens, Benoit Steiner, Ilya Sutskever, Kunal Talwar, Paul Tucker, Vincent Vanhoucke, Vijay Vasudevan, Fernanda Viégas, Oriol Vinyals, Pete Warden, Martin Wattenberg, Martin Wicke, Yuan Yu, and Xiaoqiang Zheng. TensorFlow: Large-scale machine learning on heterogeneous systems, 2015. URL http://tensorflow.org/. Software available from tensorflow.org.
    [2]
    Lee Altenberg et al. The evolution of evolvability in genetic programming. Advances in genetic programming, 3:47--74, 1994.
    [3]
    Hans-Georg Beyer and Hans-Paul Schwefel. Evolution strategies: A comprehensive introduction. Natural Computing, 1:3--52, 2002.
    [4]
    J.F.Y Brookfield. Evolution: The evolvability enigma. Current Biology, 11(3):R106 -- R108, 2001. ISSN 0960-9822.
    [5]
    Patryk Chrabaszcz, Ilya Loshchilov, and Frank Hutter. Back to basics: Benchmarking canonical evolution strategies for playing atari. arXiv preprint arXiv:1802.08842, 2018.
    [6]
    Jeff Clune, Jean-Baptiste Mouret, and Hod Lipson. The evolutionary origins of modularity. Proc. R. Soc. B, 280(1755):20122863, 2013.
    [7]
    Edoardo Conti, Vashisht Madhavan, Felipe Petroski Such, Joel Lehman, Kenneth Stanley, and Jeff Clune. Improving exploration in evolution strategies for deep reinforcement learning via a population of novelty-seeking agents. In Advances in Neural Information Processing Systems 31. 2018.
    [8]
    E Coumans and Y Bai. Pybullet, a python module for physics simulation for games, robotics and machine learning. GitHub repository, 2016.
    [9]
    Giuseppe Cuccu and Faustino Gomez. When novelty is not enough. In European Conference on the Applications of Evolutionary Computation, pages 234--243. Springer, 2011.
    [10]
    M. Ebner, M. Shackleton, and R. Shipman. How neutral networks influence evolvability. Complexity, 7(2):19--33, 2001. ISSN 1099-0526.
    [11]
    Chelsea Finn, Pieter Abbeel, and Sergey Levine. Model-agnostic meta-learning for fast adaptation of deep networks. arXiv preprint arXiv:1703.03400, 2017.
    [12]
    Dario Floreano and Francesco Mondada. Evolution of plastic neurocontrollers for situated agents. In Proc. of The Fourth International Conference on Simulation of Adaptive Behavior (SAB), From Animals to Animats. ETH Zürich, 1996.
    [13]
    Meire Fortunato, Mohammad Gheshlaghi Azar, Bilal Piot, Jacob Menick, Ian Osband, Alex Graves, Vlad Mnih, Rémi Munos, Demis Hassabis, Olivier Pietquin, Charles Blundell, and Shane Legg. Noisy networks for exploration. CoRR, abs/1706.10295, 2017. URL http://arxiv.org/abs/1706.10295.
    [14]
    Michael C. Fu. Chapter 19 gradient estimation. Simulation Handbooks in Operations Research and Management Science, 13:575, 2006.
    [15]
    J.J. Grefenstette. Evolvability in dynamic fitness landscapes: A genetic algorithm approach. In Evolutionary Computation, 1999. CEC 99. Proceedings of the 1999 Congress on, volume 3. IEEE, 2002. ISBN 0780355369.
    [16]
    Sepp Hochreiter, A Steven Younger, and Peter R Conwell. Learning to learn using gradient descent. In International Conference on Artificial Neural Networks, pages 87--94. Springer, 2001.
    [17]
    Nadav Kashtan, Elad Noor, and Uri Alon. Varying environments can speed up evolution. Proceedings of the National Academy of Sciences, 104(34):13711--13716, 2007.
    [18]
    Marc Kirschner and John Gerhart. Evolvability. Proceedings of the National Academy of Sciences, 95(15):8420--8427, 1998.
    [19]
    Loizos Kounios, Jeff Clune, Kostas Kouvaris, Günter P Wagner, Mihaela Pavlicev, Daniel M Weinreich, and Richard A Watson. Resolving the paradox of evolvability with learning theory: How evolution learns to improve evolvability on rugged fitness landscapes. arXiv preprint arXiv:1612.05955, 2016.
    [20]
    Joel Lehman and Kenneth O Stanley. Abandoning objectives: Evolution through the search for novelty alone. Evolutionary computation, 19(2):189--223, 2011.
    [21]
    Joel Lehman and Kenneth O Stanley. Improving evolvability through novelty search and self-adaptation. In IEEE Congress on Evolutionary Computation, pages 2693--2700, 2011.
    [22]
    Joel Lehman and Kenneth O Stanley. Evolving a diversity of virtual creatures through novelty search and local competition. In Proceedings of the 13th annual conference on Genetic and evolutionary computation, pages 211--218. ACM, 2011.
    [23]
    Joel Lehman and Kenneth O Stanley. Evolvability is inevitable: Increasing evolvability without the pressure to adapt. PloS one, 8(4):e62186, 2013.
    [24]
    Joel Lehman and Kenneth O Stanley. On the potential benefits of knowing everything. In Artificial Life Conference Proceedings, pages 558--565. MIT Press, 2018.
    [25]
    Geoffrey J. McLachlan, Sharon X. Lee, and Suren I. Rathnayake. Finite mixture models. Annual Review of Statistics and Its Application, 6(1):355--378, 2019.
    [26]
    Henok Mengistu, Joel Lehman, and Jeff Clune. Evolvability search. Proceedings of the 2016 on Genetic and Evolutionary Computation Conference - GECCO 16, 2016.
    [27]
    Thomas Miconi, Jeff Clune, and Kenneth O Stanley. Differentiable plasticity: training plastic neural networks with backpropagation. arXiv preprint arXiv:1804.02464, 2018.
    [28]
    Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A Rusu, Joel Veness, Marc G Bellemare, Alex Graves, Martin Riedmiller, Andreas K Fidjeland, Georg Ostrovski, et al. Human-level control through deep reinforcement learning. Nature, 518(7540):529, 2015.
    [29]
    Anh Mai Nguyen, Jason Yosinski, and Jeff Clune. Innovation engines: Automated creativity and improved stochastic optimization via deep learning. In Proceedings of the 2015 Annual Conference on Genetic and Evolutionary Computation, pages 959--966. ACM, 2015.
    [30]
    Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary DeVito, Zeming Lin, Alban Desmaison, Luca Antiga, and Adam Lerer. Automatic differentiation in PyTorch. In NIPS-W, 2017.
    [31]
    M. Pigliucci. Is evolvability evolvable? Nature Reviews Genetics, 9(1):75--82, 2008. ISSN 1471-0056.
    [32]
    Justin K Pugh, Lisa B Soros, and Kenneth O Stanley. Quality diversity: A new frontier for evolutionary computation. Frontiers in Robotics and AI, 3:40, 2016.
    [33]
    Joseph Reisinger, Kenneth O. Stanley, and Risto Miikkulainen. Towards an empirical measure of evolvability. In Genetic and Evolutionary Computation Conference (GECCO2005) Workshop Program, pages 257--264, Washington, D.C., 2005. ACM Press. URL http://nn.cs.utexas.edu/?reisinger:gecco05.
    [34]
    Tim Salimans, Jonathan Ho, Xi Chen, Szymon Sidor, and Ilya Sutskever. Evolution strategies as a scalable alternative to reinforcement learning. arXiv preprint arXiv:1703.03864, 2017.
    [35]
    John Schulman, Nicolas Heess, Theophane Weber, and Pieter Abbeel. Gradient estimation using stochastic computation graphs. CoRR, abs/1506.05254, 2015. URL http://arxiv.org/abs/1506.05254.
    [36]
    Hans-Paul Paul Schwefel. Evolution and optimum seeking: the sixth generation. John Wiley & Sons, Inc., 1993.
    [37]
    Andrea Soltoggio, John A Bullinaria, Claudio Mattiussi, Peter Dürr, and Dario Floreano. Evolutionary advantages of neuromodulated plasticity in dynamic, reward-based scenarios. In Proceedings of the 11th international conference on artificial life (Alife XI), number LIS-CONF-2008-012, pages 569--576. MIT Press, 2008.
    [38]
    Kenneth O. Stanley and Risto Miikkulainen. Evolving neural networks through augmenting topologies. Evolutionary Computation, 10:99--127, 2002.
    [39]
    Kenneth O Stanley, Bobby D Bryant, and Risto Miikkulainen. Evolving adaptive neural networks with and without adaptive synapses. In Evolutionary Computation, 2003. CEC'03. The 2003 Congress on, volume 4, pages 2557--2564. IEEE, 2003.
    [40]
    Felipe Petroski Such, Vashisht Madhavan, Edoardo Conti, Joel Lehman, Kenneth O Stanley, and Jeff Clune. Deep neuroevolution: genetic algorithms are a competitive alternative for training deep neural networks for reinforcement learning. arXiv preprint arXiv:1712.06567, 2017.
    [41]
    E. Todorov, T. Erez, and Y. Tassa. Mujoco: A physics engine for model-based control. In 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 5026--5033, Oct 2012.
    [42]
    Ricardo Vilalta and Youssef Drissi. A perspective view and survey of meta-learning. Artificial Intelligence Review, 18(2):77--95, 2002.
    [43]
    Günter P Wagner and Lee Altenberg. Perspective: complex adaptations and the evolution of evolvability. Evolution, 50(3):967--976, 1996.
    [44]
    Jane X Wang, Zeb Kurth-Nelson, Dhruva Tirumala, Hubert Soyer, Joel Z Leibo, Remi Munos, Charles Blundell, Dharshan Kumaran, and Matt Botvinick. Learning to reinforcement learn. arXiv preprint arXiv:1611.05763, 2016.
    [45]
    Daan Wierstra, Tom Schaul, Tobias Glasmachers, Yi Sun, and JÃijrgen Schmidhuber. Natural evolution strategies, 2011.
    [46]
    Bryan Wilder and Kenneth Stanley. Reconciling explanations for the evolution of evolvability. Adaptive Behavior, 23(3):171--179, 2015.

    Cited By

    View all
    • (2024)Leveraging More of Biology in Evolutionary Reinforcement LearningApplications of Evolutionary Computation10.1007/978-3-031-56855-8_6(91-114)Online publication date: 21-Mar-2024
    • (2023)Biological insights on grammar-structured mutations improve fitness and diversityProceedings of the Genetic and Evolutionary Computation Conference10.1145/3583131.3590472(558-567)Online publication date: 15-Jul-2023
    • (2023)On Evolvability and Behavior Landscapes in Neuroevolutionary Divergent SearchProceedings of the Genetic and Evolutionary Computation Conference10.1145/3583131.3590427(1203-1211)Online publication date: 15-Jul-2023
    • Show More Cited By

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    GECCO '19: Proceedings of the Genetic and Evolutionary Computation Conference
    July 2019
    1545 pages
    ISBN:9781450361118
    DOI:10.1145/3321707
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 13 July 2019

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. evolution strategy
    2. evolvability
    3. meta-learning
    4. neuroevolution

    Qualifiers

    • Research-article

    Conference

    GECCO '19
    Sponsor:
    GECCO '19: Genetic and Evolutionary Computation Conference
    July 13 - 17, 2019
    Prague, Czech Republic

    Acceptance Rates

    Overall Acceptance Rate 1,669 of 4,410 submissions, 38%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)15
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 26 Jul 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Leveraging More of Biology in Evolutionary Reinforcement LearningApplications of Evolutionary Computation10.1007/978-3-031-56855-8_6(91-114)Online publication date: 21-Mar-2024
    • (2023)Biological insights on grammar-structured mutations improve fitness and diversityProceedings of the Genetic and Evolutionary Computation Conference10.1145/3583131.3590472(558-567)Online publication date: 15-Jul-2023
    • (2023)On Evolvability and Behavior Landscapes in Neuroevolutionary Divergent SearchProceedings of the Genetic and Evolutionary Computation Conference10.1145/3583131.3590427(1203-1211)Online publication date: 15-Jul-2023
    • (2023)Instance Weighted Incremental Evolution Strategies for Reinforcement Learning in Dynamic EnvironmentsIEEE Transactions on Neural Networks and Learning Systems10.1109/TNNLS.2022.316017334:12(9742-9756)Online publication date: Dec-2023
    • (2023)Transfer Dynamics in Emergent Evolutionary CurriculaIEEE Transactions on Games10.1109/TG.2022.315102515:2(157-170)Online publication date: Jul-2023
    • (2023)Heritability of morphological and behavioural traits in evolving robotsEvolutionary Intelligence10.1007/s12065-023-00860-017:3(1733-1749)Online publication date: 3-Jul-2023
    • (2022)Simple genetic operators are universal approximators of probability distributions (and other advantages of expressive encodings)Proceedings of the Genetic and Evolutionary Computation Conference10.1145/3512290.3528746(739-748)Online publication date: 8-Jul-2022
    • (2021)Utilizing the Untapped Potential of Indirect Encoding for Neural Networks with Meta LearningApplications of Evolutionary Computation10.1007/978-3-030-72699-7_34(537-551)Online publication date: 1-Apr-2021
    • (2020)Learning to score behaviors for guided policy optimizationProceedings of the 37th International Conference on Machine Learning10.5555/3524938.3525628(7445-7454)Online publication date: 13-Jul-2020
    • (2020)Effective diversity in population based reinforcement learningProceedings of the 34th International Conference on Neural Information Processing Systems10.5555/3495724.3497239(18050-18062)Online publication date: 6-Dec-2020
    • Show More Cited By

    View Options

    Get Access

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media