Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Advertisement

Empirical study of evolutionary computation-based multi-objective Bayesian optimization for materials discovery

  • Application of soft computing
  • Published:
Soft Computing Aims and scope Submit manuscript

Abstract

Multi-objective Bayesian optimization (MOBO) is broadly used for applications with high cost observations such as materials discovery. In BO, a derivative-free optimization algorithm is generally employed to maximize the acquisition function. In this study, we present a method for acquisition function maximization based on a \((1 + 1)\)-evolutionary strategy in MOBO for materials discovery, which is a simple and easy-to-use approach with low computational complexity compared to conventional algorithms. In MOBO, weight vectors are used for scalarizing MO functions, typically employed to convert MO optimization into single-objective optimization. The weight vectors at each round of MOBO are generally obtained using either stochastic (random sampling) or deterministic methods based on searched results. To clarify the effect of both the scalarizing methods on MOBO, we examine the effectiveness of random sampling methods versus two deterministic methods: reference-vector-based and self-organizing map-based decomposition methods. Experimental results from four test functions and a hydrogen storage material database as a concrete application show the effectiveness of the proposed method and the random sampling method. These results implied that the proposed method was useful for real-world MOBO experiments in materials discovery.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Algorithm 1
Algorithm 2
Algorithm 3
Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Algorithm 4

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data availability

Enquiries about data availability should be directed to the authors.

Notes

  1. The database was constructed from the datasets obtained from http://hydrogenmaterialssearch.govtools.us (data last accessed on Apr., 2021).

  2. To improve convergence of the algorithms, the initial values were adjusted as follows. For L-BFGS-B, the initial values were 0.1 for TF1 and TF2; for CMA-ES, the initial values were 0.01 for TF1. CMA-ES was implemented by referring to https://github.com/CMA-ES/c-cmaes. L-BFGS-B was used in the SciPy (1.9.3) library in Python.

  3. For Tf2 and Tf3, there were significant differences (\( < 5 \% \) significant level) between ES and L-BFGS-B; for Tf2, there was significant difference between ES and CMA-ES; for Tf4, CMA-ES and L-BFGS-B significantly outperformed ES.

  4. For qParEGO, qEHVI, and qNEHVI, we used BoTorch (0.8.4) in Python (3.10.9), PyTorch (2.0.0) and Cuda (11.7).

  5. In the comparisons, we used the HVR metric since two state-of-the-art algorithms use the hypervolume.

  6. For F3, the state-of-the-art algorithms significantly outperformed ES. However, there were no significant differences (\( > 5 \% \) significant level) between them for F1, F2, and F4.

References

  • Akimoto Y, Nagata Y, Ono I, Kobayashi S (2012) Theoretical foundation for CMA-ES from information geometry perspective. Algorithmica 64(4):698–716

    Article  MathSciNet  Google Scholar 

  • Audet C, Bigeon J, Cartier D, Le Digabel S, Salomon L (2021) Performance indicators in multiobjective optimization. Eur J Oper Res 292(2):397–422. https://doi.org/10.1016/j.ejor.2020.11.016

    Article  MathSciNet  Google Scholar 

  • Auger A, Hansen N (2011) Theory of evolution strategies: a new perspective. In: Auger A, Doerr B (eds) Theory of randomized search heuristics: foundations and recent developments, vol 10. World Scientific Publishing, pp 289–325

    Chapter  Google Scholar 

  • Balandat M, Karrer B, Jiang D, Daulton S, Letham B, Wilson AG, Bakshy E (2020) Botorch: a framework for efficient Monte-Carlo bayesian optimization. In: Larochelle H, Ranzato M, Hadsell R, Balcan M, Lin H (eds) Advances in neural information processing systems, vol 33. Curran Associates Inc., pp 21524–21538

    Google Scholar 

  • Bechikh S, Elarbi M, Ben Said L (2017) Many-objective optimization using evolutionary algorithms: a survey. Springer, Cham, pp 105–137

    Google Scholar 

  • Bezdek JC (1981) Pattern recognition with fuzzy objective function algorithms. Kluwer Academic Publishers, New York

    Book  Google Scholar 

  • Bonilla EV, Chai KM, Williams C (2008) Multi-task Gaussian process prediction. In: Platt JC, Koller D, Singer Y, Roweis ST (eds) Advances in neural information processing systems, vol 20. Curran Associates Inc, pp 153–160

    Google Scholar 

  • Brochu E, Cora VM, de Freitas N (2010) A tutorial on Bayesian optimization of expensive cost functions, with application to active user modeling and hierarchical reinforcement learning. arXiv:1012.2599

  • Campigotto P, Passerini A, Battiti R (2014) Active learning of pareto fronts. IEEE Trans Neural Netw Learn Syst 25(3):506–519

    Article  Google Scholar 

  • Chugh T, Sindhya K, Hakanen J, Miettinen K (2019) A survey on handling computationally expensive multiobjective optimization problems with evolutionary algorithms. Soft Comput 23(9):3137–3166

    Article  Google Scholar 

  • Cubuk E, Sendek A, Reed E (2019) Screening billions of candidates for solid lithium-ion conductors: a transfer learning approach for small data. J Chem Phys 150:214701

    Article  Google Scholar 

  • Daulton S, Balandat M, Bakshy E (2020) Differentiable expected hypervolume improvement for parallel multi-objective bayesian optimization. In: Proceedings of the 34th international conference on neural information processing systems, NIPS’20. Curran Associates Inc., Red Hook

  • Daulton S, Balandat M, Bakshy E (2021) Parallel bayesian optimization of multiple noisy objectives with expected hypervolume improvement. In: Neural information processing systems

  • Deb K, Thiele L, Laumanns M, Zitzler E (2005) Scalable test problems for evolutionary multiobjective optimization. Springer, London, pp 105–145. https://doi.org/10.1007/1-84628-137-7_6

    Book  Google Scholar 

  • Frazier PI, Wang J (2016) Bayesian optimization for materials design. Springer, Cham, pp 45–75

    Google Scholar 

  • Frazier PI, Powell WB, Dayanik S (2008) A knowledge-gradient policy for sequential information collection. SIAM J Control Optim 47(5):2410–2439. https://doi.org/10.1137/070693424

    Article  MathSciNet  Google Scholar 

  • Golovin D, Zhang QR (2020) Random hypervolume scalarizations for provable multi-objective black box optimization. In: Proceedings of the 37th international conference on machine learning, ICML’20. JMLR.org

  • Gu F, Cheung Y (2018) Self-organizing map-based weight design for decomposition-based many-objective evolutionary algorithm. IEEE Trans Evol Comput 22(2):211–225

    Article  Google Scholar 

  • Gu F, Liu HL, Tan K (2012) A multiobjective evolutionary algorithm using dynamic weight design method. Int J Innov Comput Inf Control 8:3677–3688

    Google Scholar 

  • Halim AH, Ismail I, Das S (2020) Performance assessment of the metaheuristic optimization algorithms: an exhaustive review. Artif Intell Rev 54:2323–2409

    Article  Google Scholar 

  • Hansen N, Ostermeier A (2001) Completely derandomized self-adaptation in evolution strategies. Evol Comput 9(2):159–195

    Article  Google Scholar 

  • Jaszkiewicz A (2002) On the performance of multiple-objective genetic local search on the 0/1 knapsack problem-a comparative experiment. IEEE Trans Evolut Comput 6(4):402–412

    Article  Google Scholar 

  • Jones DR, Perttunen CD, Stuckman BE (1993) Lipschitzian optimization without the Lipschitz constant. J Optim Theory Appl 79(1):157–181

    Article  MathSciNet  Google Scholar 

  • Kern S, Müller SD, Hansen N, Büche D, Ocenasek J, Koumoutsakos P (2004) Learning probability distributions in continuous evolutionary algorithms—a comparative review. Nat Comput Int J 3(3):355–356

    Article  MathSciNet  Google Scholar 

  • Knowles J (2006) Parego: a hybrid algorithm with on-line landscape approximation for expensive multiobjective optimization problems. IEEE Trans Evol Comput 10(1):50–66

    Article  Google Scholar 

  • Li H, Zhang Q (2009) Multiobjective optimization problems with complicated pareto sets, MOEA/D and NSGA-II. IEEE Trans Evol Comput 13(2):284–302

    Article  Google Scholar 

  • Li B, Li J, Tang K, Yao X (2015) Many-objective evolutionary algorithms: a survey. ACM Comput Surv 48:1–35

    Article  Google Scholar 

  • Li M, Yao X (2017) What weights work for you? Adapting weights for any pareto front shape in decomposition-based evolutionary multi-objective optimisation. arXiv:1709.02679

  • Lyu W, Yang F, Yan C, Zhou D, Zeng X (2018) Multi-objective Bayesian optimization for analog/rf circuit synthesis. In: Proceedings of the 55th annual design automation conference, DAC ’18. Association for Computing Machinery, New York

  • Miettinen K (1999) Nonlinear multiobjective optimization. Kluwer Academic Publishers, Boston

    Google Scholar 

  • Mockus J, Tiesis V, Zilinskas A (2014) The application of Bayesian methods for seeking the extremum. In: Dixon LCW, Szego GP (eds) Towards global optimization, vol 2. North-Holand, pp 117–129

    Google Scholar 

  • Nikolaus H (2016) The CMA evolution strategy: a tutorial. CoRR arXiv:1604.00772

  • Paria B, Kandasamy K, Póczos B (2019) A flexible framework for multi-objective Bayesian optimization using random scalarizations. In: Proceedings of the thirty-fifth conference on uncertainty in artificial intelligence, UAI 2019, Tel Aviv, Israel, p 267

  • Qi Y, Ma X, Liu F, Jiao L, Sun J, Wu J (2014) MOEA/D with adaptive weight adjustment. Evol Comput 22(2):231–264

    Article  Google Scholar 

  • Rasmussen CE, Williams CKI (2006) Gaussian processes for machine learning. The MIT Press

    Google Scholar 

  • Riquelme N, Von Lücken C, Barán B (2015) Performance metrics in multi-objective optimization. In: 2015 Latin American computing conference (CLEI), pp 1–11. https://doi.org/10.1109/CLEI.2015.7360024

  • Schlapbach L, Züttel A (2001) Hydrogen-storage materials for mobile applications. Nature 414(6861):353–358

    Article  Google Scholar 

  • Schweidtmann AM, Clayton AD, Holmes N, Bradford E, Bourne RA, Lapkin AA (2018) Machine learning meets continuous flow chemistry: automated optimization towards the pareto front of multiple objectives. Chem Eng J 352:277–282

    Article  Google Scholar 

  • Shahriari B, Swersky K, Wang Z, Adams RP, de Freitas N (2016) Taking the human out of the loop: a review of Bayesian optimization. Proc IEEE 104(1):148–175

    Article  Google Scholar 

  • Solomou A, Zhao G, Boluki S, Joy JK, Qian X, Karaman I, Arróyave R, Lagoudas DC (2018) Multi-objective Bayesian materials discovery: application on the discovery of precipitation strengthened niti shape memory alloys through micromechanical modeling. Mater Des 160:810–827

    Article  Google Scholar 

  • Srinivas N, Krause A, Kakade SM, Seeger MW (2012) Information-theoretic regret bounds for Gaussian process optimization in the bandit setting. IEEE Trans Inf Theory 58(5):3250–3265

    Article  MathSciNet  Google Scholar 

  • Suzumura A, Ohno H, Kikkawa N, Takechi K (2022) Finding a novel electrolyte solution of lithium-ion batteries using an autonomous search system based on ensemble optimization. J Power Sources 541:231698. https://doi.org/10.1016/j.jpowsour.2022.231698

    Article  Google Scholar 

  • Turchetta M, Krause A, Trimpe S (2019) Robust model-free reinforcement learning with multi-objective Bayesian optimization. arXiv:1910.13399

  • Wang Z, Zoghi M, Hutter F, Matheson D, de Freitas N (2016) Bayesian optimization in a billion dimensions via random embeddings. J Artif Intell Res 55:361–387

    Article  MathSciNet  Google Scholar 

  • Wang Z, Gehring C, Kohli P, Jegelka S (2018) Batched large-scale Bayesian optimization in high-dimensional spaces. In: Storkey A, Perez-Cruz F (eds) Proceedings of the twenty-first international conference on artificial intelligence and statistics, PMLR, Playa Blanca, Lanzarote, Canary Islands, Proceedings of machine learning research, vol 84, pp 745–754

  • Wilson J, Hutter F, Deisenroth M (2018) Maximizing acquisition functions for Bayesian optimization. In: Bengio S, Wallach H, Larochelle H, Grauman K, Cesa-Bianchi N, Garnett R (eds) Advances in neural information processing systems, vol 31. Curran Associates Inc

    Google Scholar 

  • Wu W, He J, Tan AH, Tan CL (2002) On quantitative evaluation of clustering systems. Netw Theory Appl. https://doi.org/10.1007/978-1-4613-0227-8_4

    Article  Google Scholar 

  • Wu M, Kwong S, Jia Y, Li K, Zhang Q (2017) Adaptive weights generation for decomposition-based multi-objective optimization using Gaussian process regression. In: Proceedings of the genetic and evolutionary computation conference, GECCO ’17. Association for Computing Machinery, New York, pp 641—648

  • Yang XS (2010) Test problems in optimization. arXiv:1008.0549

  • Zhang Q, Li H (2007) MOEA/D: a multiobjective evolutionary algorithm based on decomposition. IEEE Trans Evolut Comput 11(6):712–731

    Article  Google Scholar 

  • Zhang Q, Liu W, Tsang E, Virginas B (2010) Expensive multiobjective optimization by MOEA/D with Gaussian process model. IEEE Trans Evol Comput 14(3):456–474

    Article  Google Scholar 

  • Zhang C, Tan KC, Lee LH, Gao L (2018) Adjust weight vectors in MOEA/D for bi-objective optimization problems with discontinuous pareto fronts. Soft Comput 22(12):3997–4012

  • Zitzler E, Thiele L (1999) Multiobjective evolutionary algorithms: a comparative case study and the strength pareto approach. IEEE Trans Evol Comput 3(4):257–271. https://doi.org/10.1109/4235.797969

    Article  Google Scholar 

  • Zuluaga M, Krause A, Püschel M (2016) \(\epsilon \)-PAL: an active learning approach to the multi-objective optimization problem. J Mach Learn Res 17(1):3619–3650

    MathSciNet  Google Scholar 

Download references

Acknowledgements

The author would like to thank Dr. Kazutoshi Miwa for helpful discussions about hydrogen storage materials.

Funding

The authors have not disclosed any funding.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hiroshi Ohno.

Ethics declarations

Conflict of interest

The author declares that there is no conflict of interest regarding the publication of this article.

Human and animals rights

This article does not contain any studies with human participants or animals performed by the author.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendices

Loss curves of ES and DIR

In Fig. 16, the shaded areas indicate the standard deviation. As F1 and F2 results, as increasing the noise level, the convergence of ES became slower.

Fig. 16
figure 16

Loss curves of ES and DIR. The blue line denotes ES result, and the red line denotes DIR result. The Tchebychev scalarization was used

Fig. 17
figure 17

Actual iteration number of ES with the Tchebychev scalarization for the test functions. The parentheses denote the noise level

Actual iteration number of ES

For the experiments of the test functions, the actual iteration number of ES (Algorithm 2) at each round is shown in Fig. 17. The solid line denotes the averaged iteration number, and the shaded areas indicate the standard deviation. In the sub-captions, the noise level is shown in the parentheses. As can be seen, the standard deviations were relatively large because \( M = 200 \).

Table 9 Wall time results of ES and DIR (\( T = 200 \))
Fig. 18
figure 18

Approximated PFs for F1. The red cure denotes PF (ground truth). DIR, orig-ES, and ES used the Tchebychev scalarization

Wall time comparison between ES and DIR

The wall time results (mean and standard deviation values) for the test functions are shown in Table 9. Here, we implemented the algorithms by the C language on a Linux machine with Intel Core i7-5830K processor (3.3GHz) and memory 64GB. From this table, the average speedup of ES was about \( 15 \% \), while the variation of the wall time of ES became larger than that of DIR.

Fig. 19
figure 19

Approximated PFs for F2. The red cure denotes PF (ground truth). DIR, orig-ES, and ES used the Tchebychev scalarization

Fig. 20
figure 20

Approximated PFs for F3. The red cure denotes PF (ground truth). DIR, orig-ES, and ES used the Tchebychev scalarization

Fig. 21
figure 21

Approximated PFs for F4. The red cure denotes PF (ground truth). DIR, orig-ES, and ES used the Tchebychev scalarization

Table 10 Test functions (Yang 2010)

Approximated PFs

See Figs. 18, 19, 20 and 21.

Test functions for function optimization

See Table 10.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ohno, H. Empirical study of evolutionary computation-based multi-objective Bayesian optimization for materials discovery. Soft Comput 28, 8807–8834 (2024). https://doi.org/10.1007/s00500-023-09058-z

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00500-023-09058-z

Keywords