Abstract
In this contribution, we improve of the performance of the Rectified Linear Unit Memristor Like Activation Function with the implication to help training process of CNN without a lot of epochs by computing the best value of the flatness network parameter (p). In this regards the flatness network parameter threshold (p) has been investigated and a good performance of the activation function has been discovered at p = 4.5. We firstly used the MNIST and the CIFAR-10 datasets to trained and test the Alex-Net architecture model of convolutional neural network (CNN) and we showed better performances of Rectified Linear Unit Memristor-like Activation Function compared to those of the literature. We noticed that the performance of Alex-Net also improved and the better performance was recorded when p = 4.5 with 99.50%, 99.25%, 98.81% for training, validation and testing accuracy respectively when using the MNIST. These results open the outcome to reduce the training time of neural networks when this activation function is used.







Similar content being viewed by others
References
Cerquitelli T, Meo M, Curado M, Skorin-Kapov L, Tsiropoulou EE. Machine learning empowered computer networks. Elsevier; 2023. p. 109807.
Chandrasekar A, Rakkiyappan R. Impulsive controller design for exponential synchronization of delayed stochastic memristor-based recurrent neural networks. Neurocomputing. 2016;173:1348–55. https://doi.org/10.1016/j.neucom.2015.08.088.
Chen H-C, Widodo AM, Wisnujati A, Rahaman M, Lin JC-W, Chen L, Weng C-E. AlexNet convolutional neural network for disease detection and classification of tomato leaf. Electronics. 2022;11(6):951. https://doi.org/10.3390/electronics11060951.
Clevert D-A, Unterthiner T, Hochreiter S. Fast and accurate deep network learning by exponential linear units (elus). 2015. arXiv:1511.07289. https://doi.org/10.48550/arXiv.1511.07289.
Duan F, Chapeau-Blondeau F, Abbott D. Optimized injection of noise in activation functions to improve generalization of neural networks. Chaos Solitons Fractals. 2024;178: 114363. https://doi.org/10.1016/j.chaos.2023.114363.
Elakkiya MK. Novel deep learning models with novel integrated activation functions for autism screening: AutiNet and MinAutiNet. Expert Syst Appl. 2024;238: 122102. https://doi.org/10.1016/j.eswa.2023.122102.
Emanuel RH, Docherty PD, Lunt H, Möller K. The effect of activation functions on accuracy, convergence speed, and misclassification confidence in CNN text classification: a comprehensive exploration. J Supercomput. 2024;80(1):292–312. https://doi.org/10.1007/s11227-023-05441-7.
Fernandez A, Mali A. Stable and robust deep learning by hyperbolic tangent exponential linear unit (TeLU). 2024. arXiv:2402.02790.
Gaur L, Bhatia U, Jhanjhi N, Muhammad G, Masud M. Medical image-based detection of COVID-19 using deep convolution neural networks. Multim Syst. 2023;29(3):1729–38. https://doi.org/10.1007/s00530-021-00794-6.
He K, Zhang X, Ren S, Sun J. Delving deep into rectifiers: surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE international conference on computer vision. 2015.
Karlik B, Olgac AV. Performance analysis of various activation functions in generalized MLP architectures of neural networks. Int J Artif Intell Expert Syst. 2011;1(4):111–22.
Ketwongsa W, Boonlue S, Kokaew U. A new deep learning model for the classification of poisonous and edible mushrooms based on improved AlexNet convolutional neural network. Appl Sci. 2022;12(7):3409. https://doi.org/10.3390/app12073409.
Klambauer G, Unterthiner T, Mayr A, Hochreiter S. Self-normalizing neural networks. Adv Neural Inf Process Syst. 2017;30.
Krizhevsky A, Hinton G. Learning multiple layers of features from tiny images. 2009.
Krizhevsky A, Sutskever I, Hinton GE. ImageNet classification with deep convolutional neural networks. Commun ACM. 2017;60(6):84–90. https://doi.org/10.1145/3065386.
LeCun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proc IEEE. 1998;86(11):2278–324. https://doi.org/10.1109/5.726791.
Maas AL, Hannun AY, Ng AY. Rectifier nonlinearities improve neural network acoustic models. In: Proceedings of icml. 2013.
Menaka R, Karthik R, Saranya S, Niranjan M, Kabilan S. An improved AlexNet model and cepstral coefficient-based classification of autism using EEG. Clin EEG Neurosci. 2024;55(1):43–51. https://doi.org/10.1177/15500594231178274.
Nair V, Hinton GE. Rectified linear units improve restricted boltzmann machines. In: Proceedings of the 27th international conference on machine learning (ICML-10). 2010.
Priya A, Vasudevan V. Brain tumor classification and detection via hybrid alexnet-gru based on deep learning. Biomed Signal Process Control. 2024;89: 105716. https://doi.org/10.1016/j.bspc.2023.105716.
Prodromakis T, Peh BP, Papavassiliou C, Toumazou C. A versatile memristor model with nonlinear dopant kinetics. IEEE Trans Electron Dev. 2011;58(9):3099–105. https://doi.org/10.1109/TED.2011.2158004.
Rajanand A, Singh P. ErfReLU: adaptive activation function for deep neural network. Pattern Anal Appl. 2024;27(2):68.
Rakkiyappan R, Chandrasekar A, Cao J. Passivity and passification of memristor-based recurrent neural networks with additive time-varying delays. IEEE Trans Neural Netw Learn Syst. 2014;26(9):2043–57. https://doi.org/10.1109/TNNLS.2014.2365059.
Ramachandran P, Zoph B, Le QV. Searching for activation functions. 2017. arXiv:1710.05941.
Shen S-L, Zhang N, Zhou A, Yin Z-Y. Enhancement of neural networks with an alternative activation function tanhLU. Expert Syst Appl. 2022;199: 117181.
Strukov DB, Williams RS. Exponential ionic drift: fast switching and low volatility of thin-film memristors. Appl Phys A. 2009;94(3):515–9. https://doi.org/10.1007/s00339-008-4975-3.
Subramanian B, Jeyaraj R, Ugli RAA, Kim J. APALU: a trainable, adaptive activation function for deep learning networks. 2024. arXiv:2402.08244. https://doi.org/10.48550/arXiv.2402.08244.
Sun J, Cai X, Sun F, Zhang J. Scene image classification method based on Alex-Net model. In: 2016 3rd international conference on informative and cybernetics for computational social systems (ICCSS). 2016.
Xu B, Wang N, Chen T, Li M. Empirical evaluation of rectified activations in convolutional network. 2015. arXiv:1505.00853. https://doi.org/10.48550/arXiv.1505.00853.
Yu Y, Adu K, Tashi N, Anokye P, Wang X, Ayidzoe MA. Rmaf: Relu-memristor-like activation function for deep learning. IEEE Access. 2020;8:72727–41. https://doi.org/10.1109/ACCESS.2020.2987829.
Zha J, Huang H, Liu Y. A novel window function for memristor model with application in programming analog circuits. IEEE Trans Circuits Syst II Express Briefs. 2015;63(5):423–7. https://doi.org/10.1109/TCSII.2015.2505959.
Zhang K, Wu Q, Liu A, Meng X. Can deep learning identify tomato leaf disease? Adv Multim. 2018;2018(1):6710865. https://doi.org/10.1155/2018/6710865.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
Authors, declare they have no conflict of interest regarding the publication of this article.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Tchepgoua Mbakop, M., Mboupda Pone, J.R., Chassem Kamdem, P. et al. Effects of the Flatness Network Parameter Threshold on the Performance of the Rectified Linear Unit Memristor-Like Activation Function in Deep Learning. SN COMPUT. SCI. 5, 1156 (2024). https://doi.org/10.1007/s42979-024-03507-8
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s42979-024-03507-8