Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3520304.3528919acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
poster

Dropout topology-assisted bidirectional learning particle swarm optimization for neural architecture search

Published: 19 July 2022 Publication History

Abstract

The neural architecture search (NAS) is a new high-complexity optimization problem emerging in recent years. Solving NAS is challenging for optimization algorithms due to the following two issues. Firstly, besides the network architectures, there are also network hyperparameters that need to be optimized, which causes the search space of NAS to be complex and large and poses a great challenge to the optimization ability of the optimization algorithm. Secondly, NAS is an expensive optimization problem with expensive computational time to evaluate candidates, which poses a great challenge to the search speed and convergence of the optimization algorithm. Therefore, this paper proposes a novel dropout topology-assisted bidirectional learning particle swarm optimization (DBLPSO) algorithm for NAS to tackle these two issues. Firstly, inspired by the dropout technique in deep learning, a sorting-assisted dropout-based neighbor topology is proposed to enhance the population diversity and the optimization ability of PSO. Secondly, a bidirectional learning strategy is proposed to improve search speed and accelerate PSO convergence. In the experiments, the performance of the DBLPSO algorithm is evaluated on 10 tabular benchmarks based on NAS-bench 201, NATS-bench, and HPO-bench, by comparing with four NAS algorithms that have achieved state-of-the-art results on the NAS tabular benchmarks. The experimental results show that DBLPSO can obtain great performance for NAS and is superior to those NAS algorithms in comparison.

References

[1]
Han-xiao Liu, Karen Simonyan, and Yi-Ming Yang. 2018. DARTS: Differentiable architecture search. In Proceeding of the International Conference on Learning Representations. 1--13.
[2]
Chen-xi Liu, et al. 2018. Progressive neural architecture search. In Proceedings of the European Conference on Computer Vision (ECCV). 19--34.
[3]
Jian-Yu Li, Zhi-Hui Zhan, and Jun Zhang. 2022. Evolutionary computation for expensive optimization: A survey. Machine Intelligence Research 19(2022), 3--23.
[4]
Zhi-Hui Zhan, et al. 2022. A survey on evolutionary computation for complex continuous optimization. Artificial Intelligence Review 55(2022), 59--110.
[5]
Zhi-Hui Zhan, Jian-Yu Li, Jun Zhang. 2022. Evolutionary deep learning: A survey. Neurocomputing 483(2022), 42--58.
[6]
Yu-Qiao Liu, et al. 2021. A Survey on evolutionary neural architecture search. IEEE Transactions on Neural Networks and Learning Systems (2021), 1--21.
[7]
Li-Yang Liu, Jun-Ji Wu, and Shao-Liang Meng. 2019. Analysis and improvement of neighborhood topology of particle swarm optimization. Journal of Computational Methods in Sciences and Engineering 19, 4(2019), 955--968.
[8]
Xue-Wen Xia, et al. 2020. An expanded particle swarm optimization based on multi-exemplar and forgetting ability. Information Sciences 508(2020), 105--120.
[9]
Nitish Srivastava, et al. 2014. Dropout: A simple way to prevent neural networks from overfitting. Journal of Machine Learning Research 15(2014), 1929--1958.
[10]
Jia Deng, et al. 2009. ImageNet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 248--255.
[11]
Barret Zoph and Quoc V.Le. 2017. Neural architecture search with reinforcement learning. In Proceedings of the International Conference of Learning Representation. 1--16.
[12]
Xuan-Yi Dong and Yi Yang. 2019. NAS-bench-201: Extending the scope of reproducible neural architecture search. In the Proceeding of the International Conference on Learning Representations. 1--16.
[13]
Xuan-Yi Dong, et al. 2021. Nats-bench: Benchmarking NAS algorithms for architecture topology and size. IEEE Transactions on Pattern Analysis and Machine Intelligence (2021).
[14]
Aaron Klein and Frank Hutter. 2019. Tabular benchmarks for joint architecture and hyperparameter optimization. arXiv:1905.04970 [cs.LG]
[15]
James Bergstra and Yoshua Bengio. 2012. Random search for hyper-parameter optimization. Journal of machine learning research 13, 2(2012), 281--305.
[16]
Esteban Real, et al. 2019. Regularized evolution for image classifier architecture search. In Proceedings of the AAAI Conference on Artificial Intelligence. 4780--4789.
[17]
Williams, Ronald J. 1992. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine learning 8, 3(1992), 229--256.
[18]
Stefan Falkner, Aaron Klein, and Frank Hutter. 2018. BOHB: Robust and efficient hyperparameter optimization at scale. In Proceeding of the International Conference on Machine Learning. 1437--1446.
[19]
Zhi-Hui Zhan, Zi-Jia Wang, Hu Jin, and Jun Zhang, "Adaptive distributed differential evolution," IEEE Transactions on Cybernetics 50, 11(2020), 4633--4647.
[20]
Zi-Jia Wang, Zhi-Hui Zhan, Sam Kwong, Hu Jin, and Jun Zhang, "Adaptive granularity learning distributed particle swarm optimization for large-scale optimization," IEEE Transactions on Cybernetics 51, 3(2021), 1175--1188.

Cited By

View all
  • (2024)Acceleration for Deep Reinforcement Learning using Parallel and Distributed Computing: A SurveyACM Computing Surveys10.1145/370345357:4(1-35)Online publication date: 10-Dec-2024

Index Terms

  1. Dropout topology-assisted bidirectional learning particle swarm optimization for neural architecture search

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    GECCO '22: Proceedings of the Genetic and Evolutionary Computation Conference Companion
    July 2022
    2395 pages
    ISBN:9781450392686
    DOI:10.1145/3520304
    Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 19 July 2022

    Check for updates

    Author Tags

    1. bidirectional learning strategy
    2. deep learning
    3. dropout topology
    4. dropout topology-assisted bidirectional learning PSO (DBLPSO)
    5. neural architecture search
    6. particle swarm optimization (PSO)

    Qualifiers

    • Poster

    Funding Sources

    • This work was supported in part by the National Key Research and Development Program of China under Grant 2019YFB2102102, in part by the National Natural Science Foundations of China (NSFC) under Grant 62176094 and Grant 61873097, in part by the Key-Area Research and Development of Guangdong Province under Grant 2020B010166002, in part by the Guangdong Natural Science Foundation Research Team under Grant 2018B030312003, and in part by the National Research Foundation of Korea (NRF-2021H1D3A2A01082705).

    Conference

    GECCO '22
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 1,669 of 4,410 submissions, 38%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)11
    • Downloads (Last 6 weeks)1
    Reflects downloads up to 12 Jan 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Acceleration for Deep Reinforcement Learning using Parallel and Distributed Computing: A SurveyACM Computing Surveys10.1145/370345357:4(1-35)Online publication date: 10-Dec-2024

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media