Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3331453.3362048acmotherconferencesArticle/Chapter ViewAbstractPublication PagescsaeConference Proceedingsconference-collections
research-article

EDA based Deep Neural Network Parameter Optimization

Published: 22 October 2019 Publication History

Abstract

Deep neural network has been applied in kinds of areas due to the excellent performance. The capacity of deep neural network relies on the parameter training algorithm which is always based on the gradient information. However, for deep neural network, it is harder and harder for training due to depth of the neural network and a large number of parameters. Therefore, kinds of techniques are proposed to cope with this problem. However, the training algorithm is almost based on the gradient information with inherent defect. Evolutionary algorithm has the advantage of global optimization capability, independent of gradient information etc. Estimation of distribution algorithm (EDA) is a typical evolutionary algorithm which relies on the probability model of population. Therefore, the EDA is adopted to train the weight and bias of deep neural network in his paper. However, the large scale optimization capability of EDA is limited. Thereby, an improved strategy is proposed to enhance the large scale optimization capability of EDA. In the improved scheme, a random selection strategy is carried out to select partial variables for probability modeling instead of all of the variables, in order to reduce the computing time and the probability of combination explosion. A simulation is carried out to exhibit the validity of the improved algorithm.

References

[1]
Y. LeCun, Y. Bengio and G. Hinton (2015). Deep learning, Nature, 521, 436--444.
[2]
G. E. Hinton and R. R. Salakhutdinov (2006). Reducing the dimensionality of data with neural networks. Science, 313, 504--507.
[3]
F. A. W. M. Mohr (2018). ML-Plan: Automated machine learning via hierarchical planning. Machine Learning, 107, 1495--1515.
[4]
M. Wistuba, A. Rawat and T. Pedapati (2019). A Survey on Neural Architecture Search. arXiv e-prints, 1905.01392.
[5]
W. Dong, T. Chen, P. Tiňo, and X. Yao (2013). Scaling Up Estimation of Distribution Algorithms for Continuous Optimization. IEEE Transactions on Evolutionary Computation, 17, 797--822.
[6]
Y. Qiu, W. Zhou, N. Yu, and P. Du (2018). Denoising Sparse Autoencoder Based Ictal EEG Classification. IEEE Transactions on Neural Systems and Rehabilitation Engineering, 1--1.
[7]
Y. Bengio, A. Courville and P. Vincent (2013). Representation learning: A review and new perspectives. IEEE transactions on pattern analysis and machine intelligence, 35, 1798--1828.
[8]
M. Hauschild and M. Pelikan (2011). An introduction and survey of estimation of distribution algorithms. Swarm and Evolutionary Computation, 1, 111--128.
[9]
Q. Xu, C. Zhang, J. Sun, and L. Zhang (2016). Adaptive Learning Rate Elitism Estimation of Distribution Algorithm Combining Chaos Perturbation for Large Scale Optimization. Open Cybernetics & Systemics Journal, 10, 20--40.
[10]
Q. Xu, C. Zhang and L. Zhang (2014). A Fast Elitism Gaussian Estimation of Distribution Algorithm and Application for PID Optimization. The Scientific World Journal, 2014, 1--14.
[11]
W. Dong, T. Chen, P. Tiňo, X. Yao (2013). Scaling Up Estimation of Distribution Algorithms for Continuous Optimization, IEEE Transactions on Evolutionary Computation 17, 797--822.
[12]
A. Kabán, J. Bootkrajang and R. J. Durrant (2016). Toward Large-Scale Continuous EDA: A Random Matrix Theory Perspective. Evolutionary Computation, 24, 255--291.
[13]
P. Vincent, H. Larochelle, I. Lajoie, Y. Bengio, P. Manzagol (2010). Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion, The Journal of Machine Learning Research 11, 3371--3408.

Index Terms

  1. EDA based Deep Neural Network Parameter Optimization

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    CSAE '19: Proceedings of the 3rd International Conference on Computer Science and Application Engineering
    October 2019
    942 pages
    ISBN:9781450362948
    DOI:10.1145/3331453
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 22 October 2019

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Deep neural network
    2. EDA
    3. Optimization
    4. Training

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Funding Sources

    Conference

    CSAE 2019

    Acceptance Rates

    Overall Acceptance Rate 368 of 770 submissions, 48%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • 0
      Total Citations
    • 122
      Total Downloads
    • Downloads (Last 12 months)10
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 18 Nov 2024

    Other Metrics

    Citations

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media