Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3377930.3390232acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
research-article

Covariance matrix adaptation for the rapid illumination of behavior space

Published: 26 June 2020 Publication History

Abstract

We focus on the challenge of finding a diverse collection of quality solutions on complex continuous domains. While quality diversity (QD) algorithms like Novelty Search with Local Competition (NSLC) and MAP-Elites are designed to generate a diverse range of solutions, these algorithms require a large number of evaluations for exploration of continuous spaces. Meanwhile, variants of the Covariance Matrix Adaptation Evolution Strategy (CMA-ES) are among the best-performing derivative-free optimizers in single-objective continuous domains. This paper proposes a new QD algorithm called Covariance Matrix Adaptation MAP-Elites (CMA-ME). Our new algorithm combines the self-adaptation techniques of CMA-ES with archiving and mapping techniques for maintaining diversity in QD. Results from experiments based on standard continuous optimization benchmarks show that CMA-ME finds better-quality solutions than MAP-Elites; similarly, results on the strategic game Hearthstone show that CMA-ME finds both a higher overall quality and broader diversity of strategies than both CMA-ES and MAP-Elites. Overall, CMA-ME more than doubles the performance of MAP-Elites using standard QD performance metrics. These results suggest that QD algorithms augmented by operators from state-of-the-art optimization algorithms can yield high-performing methods for simultaneously exploring and optimizing continuous search spaces, with significant applications to design, testing, and reinforcement learning among other domains.

Supplementary Material

ZIP File (p94-fontaine-suppl.zip)
Supplemental material.

References

[1]
Alberto Alvarez, Steve Dahlskog, Jose Font, and Julian Togelius. 2019. Empowering Quality Diversity in Dungeon Design with Interactive Constrained MAP-Elites. IEEE Conference on Games (CoG).
[2]
Kai Arulkumaran, Antoine Cully, and Julian Togelius. 2019. AlphaStar: An Evolutionary Computation Perspective. In GECCO '19: Proceedings of the Genetic and Evolutionary Computation Conference Companion, Manuel López-Ibáñez (Ed.). ACM, New York, NY, USA.
[3]
Marc G Bellemare, Yavar Naddaf, Joel Veness, and Michael Bowling. 2013. The arcade learning environment: An evaluation platform for general agents. Journal of Artificial Intelligence Research 47 (2013), 253--279.
[4]
Aditya Bhatt, Scott Lee, Fernando de Mesentier Silva, Connor W. Watson, Julian Togelius, and Amy K. Hoover. 2018. Exploring the Hearthstone Deck Space. In Proceedings of the 13th International Conference on the Foundations of Digital Games. ACM, 18.
[5]
Edoardo Conti, Vashisht Madhavan, Felipe Petroski Such, Joel Lehman, Kenneth O. Stanley, and Jeff Clune. 2018. Improving Exploration in Evolution Strategies for Deep Reinforcement Learning via a Population of Novelty-Seeking Agents. In Proceedings of the 32Nd International Conference on Neural Information Processing Systems (NIPS'18). Curran Associates Inc., USA, 5032--5043. http://dl.acm.org/citation.cfm?id=3327345.3327410
[6]
Giuseppe Cuccu, Julian Togelius, and Philippe Cudré-Mauroux. 2019. Playing atari with six neurons. In Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems. International Foundation for Autonomous Agents and Multiagent Systems, 998--1006.
[7]
Antoine Cully. 2019. Autonomous Skill Discovery with Quality-Diversity and Unsupervised Descriptors. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO '19). ACM, 81--89.
[8]
Antoine Cully, Jeff Clune, Danesh Tarapore, and Jean-Baptiste Mouret. 2015. Robots that can adapt like animals. Nature 521, 7553 (2015), 503.
[9]
Antoine Cully and Jean-Baptiste Mouret. 2013. Behavioral Repertoire Learning in Robotics. In Proceedings of the 15th Annual Conference on Genetic and Evolutionary Computation (GECCO '13). ACM, 175----182.
[10]
Antoine Cully and Jean-Baptiste Mouret. 2016. Evolving a Behavioral Repertoire for a Walking Robot. Evolutionary Computation 24 (2016), 59--88. Issue 1.
[11]
Fernando de Mesentier Silva, Rodrigo Canaan, Scott Lee, Matthew C Fontaine, Julian Togelius, and Amy K Hoover. 2019. Evolving the hearthstone meta. In 2019 IEEE Conference on Games (CoG). IEEE, 1--8.
[12]
Cedric Decoster, Jean Seong Bjorn Choe, et al. 2019. Sabberstone. https://github.com/HearthSim/SabberStone Accessed: 2019-11-01.
[13]
A. E. Eiben, R. Hinterding, and Z. Michalewicz. 1999. Parameter control in evolutionary algorithms. IEEE Transactions on Evolutionary Computation 3, 2 (July 1999), 124--141.
[14]
Blizzard Entertainment. [n. d.]. Hearthstone. https://playhearthstone.com/en-us/. ([n. d.]). Accessed: 2019-11-01.
[15]
Stefano Fioravanzo and Giovanni Iacca. 2019. Evaluating MAP-Elites on Constrained Optimization Problems. (2019). arXiv:1902.00703
[16]
Matthew Fontaine. 2019. EvoStone. https://github.com/tehqin/EvoStone Accessed: 2019-12-01.
[17]
Matthew Fontaine. 2019. QualDivBenchmark. https://github.com/tehqin/QualDivBenchmark Accessed: 2019-12-01.
[18]
Matthew C. Fontaine, Scott Lee, L. B. Soros, Fernando de Mesentier Silva, Julian Togelius, and Amy K. Hoover. 2019. Mapping Hearthstone Deck Spaces through MAP-Elites with Sliding Boundaries. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO '19). ACM, New York, NY, USA, 161--169.
[19]
Adam Gaier, Alexander Asteroth, and Jean-Baptiste Mouret. 2019. Are Quality Diversity Algorithms Better at Generating Stepping Stones than Objective-Based Search?. In Proceedings of the Genetic and Evolutionary Computation Conference Companion. ACM, 115--116.
[20]
Pablo García-Sánchez, Alberto Tonda, Antonio J Fernández-Leiva, and Carlos Cotta. 2019. Optimizing Hearthstone agents using an evolutionary algorithm. Knowledge-Based Systems (2019), 105032.
[21]
Tobias Glasmachers, Tom Schaul, Sun Yi, Daan Wierstra, and Jürgen Schmidhuber. 2010. Exponential natural evolution strategies. In Proceedings of the 12th annual conference on Genetic and evolutionary computation. ACM, 393--400.
[22]
Daniele Gravina, Ahmed Khalifa, Antonios Liapis, Julian Togelius, and Georgios Yannakakis. 2019. Procedural Content Generation through Quality Diversity. IEEE Conference on Games (CoG), 1--8.
[23]
Nikolaus Hansen. 2016. The CMA evolution strategy: A tutorial. arXiv preprint arXiv:1604.00772 (2016).
[24]
N. Hansen, A. Auger, O. Mersmann, T. Tusar, and D. Brockhof. 2016. COCO: A platform for comparing continuous optimizers in a black-box setting. (2016).
[25]
Nikolaus Hansen, Anne Auger, Raymond Ros, Steffen Finck, and Petr Pošík. 2010. Comparing Results of 31 Algorithms from the Black-Box Optimization Benchmarking BBOB-2009. Proceedings of the 12th Annual Genetic and Evolutionary Computation Conference, GECCO '10 - Companion Publication, 1689--1696.
[26]
Nikolaus Hansen and Stefan Kern. 2004. Evaluating the CMA evolution strategy on multimodal test functions. In International Conference on Parallel Problem Solving from Nature. Springer, 282--291.
[27]
Nikolaus Hansen and Andreas Ostermeier. 2001. Completely derandomized self-adaptation in evolution strategies. Evolutionary computation 9, 2 (2001), 159--195.
[28]
Amy K Hoover, Julian Togelius, Scott Lee, and Fernando de Mesentier Silva. 2019. The Many AI Challenges of Hearthstone. KI-Künstliche Intelligenz (2019), 1--11.
[29]
HSReplay. [n. d.]. HSReplay. https://hsreplay.net/. ([n. d.]). Accessed: 2019-11-01.
[30]
Ahmed Khalifa, Michael Cerny Green, Gabriella Barros, and Julian Togelius. 2019. Intentional computational level design. In Proceedings of The Genetic and Evolutionary Computation Conference. 796--803.
[31]
Ahmed Khalifa, Scott Lee, Andy Nealen, and Julian Togelius. 2018. Talakat: Bullet Hell Generation Through Constrained Map-elites. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO '18). ACM, New York, NY, USA, 1047--1054.
[32]
Joel Lehman and Kenneth O. Stanley. 2008. Exploiting Open-Endedness to Solve Problems through the Search for Novelty. In Proceedings of the Eleventh International Conference on Artificial Life (Alife XI). 329--336.
[33]
Joel Lehman and Kenneth O. Stanley. 2011. Abandoning Objectives: Evolution through the Search for Novelty Alone. Evolutionary Computation (2011).
[34]
Joel Lehman and Kenneth O. Stanley. 2011. Evolving a Diversity of Virtual Creatures through Novelty Search and Local Competition. In Proceedings of the 13th Annual Conference on Genetic and Evolutionary Computation (GECCO '11). ACM, 211--218.
[35]
Fei Liang. [n. d.]. Tempo Rogue. https://www.hearthstonetopdecks.com/decks/tempo-rogue-rise-of-shadows-1-legend-etc/. ([n. d.]). Accessed: 2019-11-01.
[36]
Andrew Y. Ng, Daishi Harada, and Stuart J. Russell. 1999. Policy Invariance Under Reward Transformations: Theory and Application to Reward Shaping. In Proceedings of the Sixteenth International Conference on Machine Learning (ICML '99). Morgan Kaufmann Publishers Inc., San Francisco, CA, USA, 278--287. http://dl.acm.org/citation.cfm?id=645528.657613
[37]
Jørgen Nordmoen, Kai Olav Ellefsen, and Kyrre Glette. 2018. Combining MAP-Elites and Incremental Evolution to Generate Gaits for a Mammalian Quadruped Robot. 719--733.
[38]
Jørgen Nordmoen, Eivind Samuelsen, Kai Olav Ellefsen, and Kyrre Glette. 2018. Dynamic Mutation in MAP-Elites for Robotic Repertoire Generation. In Artificial Life Conference Proceedings. MIT Press, 598--605.
[39]
Mike Preuss. 2012. Improved Topological Niching for Real-valued Global Optimization. In Proceedings of the 2012T European Conference on Applications of Evolutionary Computation (EvoApplications'12). Springer-Verlag, Berlin, Heidelberg, 386--395.
[40]
Justin K. Pugh, Lisa B. Soros, and Kenneth O. Stanley. 2016. Quality Diversity: A New Frontier for Evolutionary Computation. Frontiers in Robotics and AI 3 (2016), 40.
[41]
Justin K. Pugh, L. B. Soros, Paul A. Szerlip, and Kenneth O. Stanley. 2015. Confronting the Challenge of Quality Diversity. In Proceedings of the 2015 Annual Conference on Genetic and Evolutionary Computation (GECCO '15). ACM, New York, NY, USA, 967--974.
[42]
Ofer M. Shir and Thomas Bäck. 2006. Niche Radius Adaptation in the CMA-ES Niching Algorithm. In Proceedings of the 9th International Conference on Parallel Problem Solving from Nature (PPSN'06). Springer-Verlag, Berlin, Heidelberg, 142--151.
[43]
Davy Smith, Laurissa Tokarchuk, and Geraint Wiggins. 2016. Rapid phenotypic landscape exploration through hierarchical spatial partitioning. In International conference on parallel problem solving from nature. Springer, 911--920.
[44]
Maciej Świechowski, Tomasz Tajmajer, and Andrzej Janusz. 2018. Improving Hearthstone AI by Combining MCTS and Supervised Learning Algorithms. In 2018 IEEE Conference on Computational Intelligence and Games (CIG). IEEE, 1--8.
[45]
Vassilis Vassiliades, Konstantinos Chatzilygeroudis, and Jean-Baptiste Mouret. 2018. Using Centroidal Voronoi Tessellations to Scale Up the Multidimensional Archive of Phenotypic Elites Algorithm. IEEE Transactions on Evolutionary Computation 22, 4 (2018), 623--630.
[46]
Vassiiis Vassiliades and Jean-Baptiste Mouret. 2018. Discovering the elite hypervolume by leveraging interspecies correlation. In Proceedings of the Genetic and Evolutionary Computation Conference. 149--156.
[47]
Oriol Vinyals, Igor Babuschkin, Wojciech M. Czarnecki, Michaël Mathieu, Andrew Dudzik, Junyoung Chung, David H. Choi, Richard Powell, Timo Ewalds, Petko Georgiev, et al. 2019. Grandmaster level in StarCraft II using multi-agent reinforcement learning. Nature 575, 11 (2019), 350--354.

Cited By

View all
  • (2024)Multi-Agent Diagnostics for Robustness via Illuminated DiversityProceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems10.5555/3635637.3663024(1630-1644)Online publication date: 6-May-2024
  • (2024)On the Use of Quality Diversity Algorithms for the Travelling Thief ProblemACM Transactions on Evolutionary Learning and Optimization10.1145/3641109Online publication date: 17-Jan-2024
  • (2024)Summary of "Curiosity creates Diversity in Policy Search"Proceedings of the Genetic and Evolutionary Computation Conference Companion10.1145/3638530.3664076(43-44)Online publication date: 14-Jul-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
GECCO '20: Proceedings of the 2020 Genetic and Evolutionary Computation Conference
June 2020
1349 pages
ISBN:9781450371285
DOI:10.1145/3377930
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 26 June 2020

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. MAP-Elites
  2. evolutionary algorithms
  3. hearthstone
  4. illumination algorithms
  5. optimization
  6. quality diversity

Qualifiers

  • Research-article

Conference

GECCO '20
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,669 of 4,410 submissions, 38%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)150
  • Downloads (Last 6 weeks)25
Reflects downloads up to 23 Dec 2024

Other Metrics

Citations

Cited By

View all
  • (2024)Multi-Agent Diagnostics for Robustness via Illuminated DiversityProceedings of the 23rd International Conference on Autonomous Agents and Multiagent Systems10.5555/3635637.3663024(1630-1644)Online publication date: 6-May-2024
  • (2024)On the Use of Quality Diversity Algorithms for the Travelling Thief ProblemACM Transactions on Evolutionary Learning and Optimization10.1145/3641109Online publication date: 17-Jan-2024
  • (2024)Summary of "Curiosity creates Diversity in Policy Search"Proceedings of the Genetic and Evolutionary Computation Conference Companion10.1145/3638530.3664076(43-44)Online publication date: 14-Jul-2024
  • (2024)Generating Diverse Critics for Conditioned Policy DistillationProceedings of the Genetic and Evolutionary Computation Conference Companion10.1145/3638530.3654429(167-170)Online publication date: 14-Jul-2024
  • (2024)Informed Diversity Search for Learning in Asymmetric Multiagent SystemsProceedings of the Genetic and Evolutionary Computation Conference10.1145/3638529.3654206(313-321)Online publication date: 14-Jul-2024
  • (2024)Enhancing MAP-Elites with Multiple Parallel Evolution StrategiesProceedings of the Genetic and Evolutionary Computation Conference10.1145/3638529.3654089(1082-1090)Online publication date: 14-Jul-2024
  • (2024)Quality with Just Enough Diversity in Evolutionary Policy SearchProceedings of the Genetic and Evolutionary Computation Conference10.1145/3638529.3654047(105-113)Online publication date: 14-Jul-2024
  • (2024)Density Descent for Diversity OptimizationProceedings of the Genetic and Evolutionary Computation Conference10.1145/3638529.3654001(674-682)Online publication date: 14-Jul-2024
  • (2024)Parametric-Task MAP-ElitesProceedings of the Genetic and Evolutionary Computation Conference10.1145/3638529.3653993(68-77)Online publication date: 14-Jul-2024
  • (2024)Preference-Learning Emitters for Mixed-Initiative Quality-Diversity AlgorithmsIEEE Transactions on Games10.1109/TG.2023.326445716:2(303-316)Online publication date: Jun-2024
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media