Skip to main content

Evolving Artificial Neural Networks Using Butterfly Optimization Algorithm for Data Classification

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 11953))

Abstract

One of the most difficult challenges in machine learning is the training process of artificial neural networks, which is mainly concerned with determining the best set of weights and biases. Gradient descent techniques are known as the most popular training algorithms. However, they are susceptible to local optima and slow convergence in training. Therefore, several stochastic optimization algorithms have been proposed in the literature to alleviate the shortcomings of gradient descent approaches. The butterfly optimization algorithm (BOA) is a recently proposed meta-heuristic approach. Its inspiration is based on the food foraging behavior of butterflies in the nature. Moreover, it has been shown that BOA is effective in undertaking a wide range of optimization problems and attaining the global optima solutions. In this paper, a new classification method based on the combination of artificial neural networks and BOA algorithm is proposed. To this end, BOA is applied as a new training strategy by optimizing the weights and biases of artificial neural networks. This leads to improving the convergence speed and also reducing the risk of falling into local optima. The proposed classification method is compared with other state-of-the-art methods based on two well-known data sets and different evaluation measures. The experimental results ascertain the superiority of the proposed method in comparison with the other methods.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Agrawal, S., Agrawal, J.: Neural network techniques for cancer prediction: a survey. Procedia Comput. Sci. 60, 769–774 (2015)

    Article  Google Scholar 

  2. Ahmadian, S., Khanteymoori, A.R.: Training back propagation neural networks using asexual reproduction optimization. In: 2015 7th Conference on Information and Knowledge Technology (IKT), pp. 1–6. IEEE (2015)

    Google Scholar 

  3. Aljarah, I., et al.: Optimizing connection weights in neural networks using the whale optimization algorithm. Soft Comput. 22(1), 1–15 (2016)

    Article  Google Scholar 

  4. Amiri, M.J., Abedi-Koupai, J., Jalali, S.M.J., Mousavi, S.F.: Modeling of fixed-bed column system of Hg(II) ions on ostrich bone Ash/nZVI composite by artificial neural network. J. Environ. Eng. 143(9), 04017061 (2017)

    Article  Google Scholar 

  5. Arora, S., Singh, S.: Butterfly optimization algorithm: a novel approach for global optimization. Soft. Comput. 23(3), 715–734 (2019)

    Article  Google Scholar 

  6. Chau, K.W.: Particle swarm optimization training algorithm for ANNs in stage prediction of Shing Mun River. J. Hydrol. 329(3–4), 363–367 (2006)

    Article  Google Scholar 

  7. Dheeru, D., Taniskidou, E.K.: UCI machine learning repository (2017)

    Google Scholar 

  8. Faris, H., Aljarah, I., Mirjalili, S.: Training feedforward neural networks using multi-verse optimizer for binary classification problems. Appl. Intell. 45(2), 322–332 (2016)

    Article  Google Scholar 

  9. Faris, H., Aljarah, I., Mirjalili, S.: Improved monarch butterfly optimization for unconstrained global search and neural network training. Appl. Intell. 48(2), 445–464 (2018)

    Article  Google Scholar 

  10. Gupta, J.N., Sexton, R.S.: Comparing backpropagation with a genetic algorithm for neural network training. Omega 27(6), 679–684 (1999)

    Article  Google Scholar 

  11. Jalali, S.M.J., Park, H.W.: Conversations about open data on twitter. Int. J. Contents 13(1), 31–37 (2017)

    Google Scholar 

  12. Jalali, S.M.J., Karimi, M., Khosravi, A., Nahavandi, S.: An efficient neuroevolution approach for heart disease detection. In: 2019 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 1–6 (2019)

    Google Scholar 

  13. Jalali, S.M.J., Khosravi, A., Kebria, P.M., Hedjam, R., Nahavandi, S.: Autonomous robot navigation system using the evolutionary multi-verse optimizer algorithm. In: 2019 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 1–6 (2019)

    Google Scholar 

  14. Jalali, S.M.J., Mahdizadeh, E., Mahmoudi, M.R., Moro, S.: Analytical assessment process of e-learning domain research between 1980 and 2014. Int. J. Manag. Educ. 12(1), 43–56 (2018)

    Article  Google Scholar 

  15. Jalali, S.M.J., Park, H.W.: State of the art in business analytics: themes and collaborations. Qual. Quant. 52(2), 627–633 (2018)

    Article  Google Scholar 

  16. Jalali, S.M.J., Moro, S., Mahmoudi, M.R., Ghaffary, K.A., Maleki, M., Alidoostan, A.: A comparative analysis of classifiers in cancer prediction using multiple data mining techniques. Int. J. Bus. Intell. Syst. Eng. 1(2), 166–178 (2017)

    Google Scholar 

  17. Mafarja, M., et al.: Binary grasshopper optimisation algorithm approaches for feature selection problems. Expert Syst. Appl. 117, 267–286 (2019)

    Article  Google Scholar 

  18. Mirjalili, S.: How effective is the grey wolf optimizer in training multi-layer perceptrons. Appl. Intell. 43(1), 150–161 (2015)

    Article  Google Scholar 

  19. Mirjalili, S., Hashim, S.Z.M., Sardroudi, H.M.: Training feedforward neural networks using hybrid particle swarm optimization and gravitational search algorithm. Appl. Math. Comput. 218(22), 11125–11137 (2012)

    MathSciNet  MATH  Google Scholar 

  20. Mirjalili, S., Mirjalili, S.M., Lewis, A.: Let a biogeography-based optimizer train your multi-layer perceptron. Inf. Sci. 269, 188–209 (2014)

    Article  MathSciNet  Google Scholar 

  21. Shahzadeh, A., Khosravi, A., Nahavandi, S.: Improving load forecast accuracy by clustering consumers using smart meter data. In: 2015 International Joint Conference on Neural Networks (IJCNN), pp. 1–7. IEEE (2015)

    Google Scholar 

  22. She, F.H., Kong, L.X., Nahavandi, S., Kouzani, A.Z.: Intelligent animal fiber classification with artificial neural networks. Text. Res. J. 72(7), 594–600 (2002)

    Article  Google Scholar 

  23. Shehu, G.S., Çetinkaya, N.: Flower pollination-feedforward neural network for load flow forecasting in smart distribution grid. Neural Comput. Appl. 31(10), 1–12 (2018)

    Google Scholar 

  24. Vanani, I.R., Jalali, S.M.J.: Analytical evaluation of emerging scientific trends in business intelligence through the utilisation of burst detection algorithm. Int. J. Bibliometr. Bus. Manag. 1(1), 70–79 (2017)

    Article  Google Scholar 

  25. Vanani, I.R., Jalali, S.M.J.: A comparative analysis of emerging scientific themes in business analytics. Int. J. Bus. Inf. Syst. 29(2), 183–206 (2018)

    Google Scholar 

  26. Wolpert, D.H., Macready, W.G.: No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1(1), 67–82 (1997)

    Article  Google Scholar 

  27. Zhang, Q., Yang, L.T., Chen, Z., Li, P.: A survey on deep learning for big data. Inf. Fusion 42, 146–157 (2018)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Seyed Mohammad Jafar Jalali .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Jalali, S.M.J., Ahmadian, S., Kebria, P.M., Khosravi, A., Lim, C.P., Nahavandi, S. (2019). Evolving Artificial Neural Networks Using Butterfly Optimization Algorithm for Data Classification. In: Gedeon, T., Wong, K., Lee, M. (eds) Neural Information Processing. ICONIP 2019. Lecture Notes in Computer Science(), vol 11953. Springer, Cham. https://doi.org/10.1007/978-3-030-36708-4_49

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-36708-4_49

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-36707-7

  • Online ISBN: 978-3-030-36708-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics