1 Introduction

Classification is a form of supervised machine learning that is commonly used to support the decision-making procedure in different areas such as science, medicine, business and industry. Classification challenges appear when an object has to be classified based on a number of object attributes [1]. The most crucial step in the process of classification is to select the right classification technique that can relate to all real-world problems. The classification techniques are open for use only after the completion of primary testing and after the results have been deemed as acceptable [2]. Several techniques have been formulated by researchers based on artificial intelligence, such as artificial neural networks (ANNs) [3], naive Bayes classifier [4], support vector machines [5], radial basis function network [6], K-nearest neighbours algorithm, iterative dichotomiser 3 (ID3) algorithm [7], and many more [8,9,10].

For solving the class statistics, the NNs use heuristic models, which iteratively alter the system parameters so that the system performance can be improved. On the other hand, this technique expends a great deal of computation time during training, which can result in false minima. To disable or minimise this shortcoming of the methodology, the probabilistic NN (PNN) was established, which is a classification technique that depends on statistical principles [11, 12].

The PNN refers to a feed-forward NN that can be applied to solve non-linear problems. This is particularly possible because in PNN, the sigmoid activation function of the NN is substituted by an exponential function that is derived from a Bayesian decision rule [13]. The structure of the PNN constitutes of four layers: the input layer, the pattern layer, the summation layer, and finally the output layer [14]. A PNN has the potential to attain optimal classification, and is more accurate than NNs. However, a PNN has its limitations—for example, it is slower than an NN at classifying new objects or instances, and it requires additional computational resources, such as memory space [15, 16].

Different classification techniques have been hybridised by researchers, where the positives or advantages of the individual algorithms have been combined so that improved performance can be achieved. Therefore, the limitations of the classification techniques have been minimised to an extent with the help of metaheuristic algorithms, and thus efficiency of the classification techniques has been enhanced. In general, researchers have made use of simulations of different aspects of the natural world in order to solve various optimisation problems [17]. The primary objective behind integrating two algorithms is to combine the strengths of the algorithms and overcome their individual limitations. Every metaheuristic algorithm mentioned here has its own successful domain. For instance, the GA algorithm is very effective with regard to exploration, but is not useful in exploitation [18, 19].

Various metaheuristic algorithms have been combined with NNs to explore the search space effectively and efficiently [20]. According to the study of combination of GAs and ANNs, it has been established that GAs can optimise NNs [21,22,23,24,25,26]. Moreover, a differential evolution algorithm has been applied in order to obtain the best parameters for an NN [27, 28]. The ant colony optimisation [29,30,31,32] is an extension of this. A hybrid particle swarm optimisation (PSO) algorithm has also been used to optimise the training of NNs [33,34,35,36,37,38,39,40,41,42]. The limitation of the back propagation (BP) technique has been overcome with the help of the harmony search algorithms [43,44,45]. Another feed-forward network training that has been used is the cuckoo search [46,47,48]. This has been combined with the firefly algorithm (FA) with NN [49, 50], and lastly, the biogeography-based optimisation technique has been applied [51].

In this study, we analyse the performance of the PNN in solving classification problems with the help of the water cycle algorithm (WCA) metaheuristic algorithm [52]. In the recommended model, the PNN creates the initial solution arbitrarily and the performance enhancement progresses with the WCA that improves the weights of the PNN. That is, the potential of using the search ability of the WCA in order to increase the performance of the PNN is analysed. Further, it is analysed how to achieve this by exploiting and exploring the search space more effectively and by regulating the random steps. Finally, it is assessed how the WCA can avoid premature convergence and immobility of the population so that the PNN classification technique can find the optimal solution.

The remaining paper is organised as follows. Section 2 provides the backdrop and literature review. Section 3 emphasises on the recommended model. The experiments, outcomes and discussions are presented in Sect. 4. Lastly, Sect. 5 presents a conclusion and offers an insight for future research.

2 Background and literature: water cycle algorithm (WCA)

WCA is a new metaheuristic technique that has been implemented in a number of constrained engineering design problems and it was formulated by Eskandar, Sadollah et al. (2012). Numerous studies have indicated that the WCA is more effectual than other popular optimisers with respect to accuracy and convergence speed [52].

Eskandar, Sadollah et al. (2013) proposed a technique for the optimisation of truss structures by weight sizing as well as discrete and continues variables with the use of WCA. There was then a comparison of the results with the results of other efficient optimisers. They found out that the WCA optimiser was able to give optimal solutions and better convergence speed rate [53].

Haddad, Moravej et al. (2014) proposed using WCA to determine optimal approaches that can be used for reservoir systems. Evolutionary algorithms are dependable methods that can be used to difficult problems. The results offer the best reliability and efficacy in answering reservoir process problems [54].

Jabbar and Zainudin (2014) proposed the utilisation of the WCA as a mathematical tool for reducing an attribute by evaluating the quality of solutions. The experiments were evidence that the new technique is capable of accomplishing higher performance compared to the other attribute selection methods [55].

Sadollah, Eskandar et al. (2015) proposed solving constrained problems using the multiobjective WCA (MOWCA). They then proposed using MOWCA to explore sets of non-dominated solutions. They then compared the results to the result of other efficient methods using tabular, descriptive, and graphical presentations. They discovered that the proposed method exhibited better performance [56].

Sadollah, Eskandar et al. (2015) suggested a new methodology for the enhancement of the evaporation rate that relies on WCA for search improvement. A comparison between the standard WCA and ER-WCA was then conducted and it was discovered that ER-WCA exhibited a better balance between exploration and exploitation. Furthermore, the researchers observed that it was better at finding all the global optima. Additionally, the results revealed that the ER-WCA achieved a higher convergence speed for the global solution and better accuracy compared to other optimisers and the standard WCA [57].

Sadollah, Eskandar et al. (2015) proposed the utilisation of WCA in answering multiobjective optimization problems (MOPs). The researchers utilised several benchmark problems to assess the proposed WCA’s performance. The results proved the WCA’s efficacy in terms of its ability to solve MOPs and its exploration capability [58].

Sarvi and Avanaki (2015) introduced a WCA that could be useful in reducing power supply loss and maintaining and operating the power supply of batteries. It was found that the proposed method performed better in maintenance, operation and minimisation of the loss of power supply probability. Furthermore, it optimised the battery’s charge level, which enhanced battery life [59].

El-Hameed and El-Fergany (2016) suggested a methodology that was based on using WCA as the constrained optimiser in finding the optimal parameters so that the interconnected power systems can be protected against random load. The dynamic performances of the proposed WCA methodology was discovered to lead to improved performance of the optimised parameters under varying sampling times [60].

Khalilpourazari and Mohammadi (2016) proposed the utilisation of a WCA metaheuristic in solving the mathematical model for a closed-loop supply chain. The proposed design was used to identify parameters such as material flow, locations, and vehicle numbers in order to reduce costs of the total supply chain. To evaluate WCA efficiency, several test problems were effectively answered [61].

Sadollah, Eskandar et al. (2016) gave a detailed open-source code for the WCA to elucidate the manner by which the algorithm effectively performs in solving unconstrained and constrained optimisation problems [62].

Heidari, Abbaspour et al. (2017) suggested a new WCA process to improve the conventional algorithm’s performance in finding the best classification solution. The WCA implemented and modified the chaotic signal functions to obtain the best signal. Then, the chaotic algorithm trained an NN to tackle benchmark problems. According to the statistical results, the chaotic WCA with sinusoidal map was efficient in exploiting high-quality solutions and was able to perform better than the other investigated algorithms [63].

Méndez, Castillo et al. (2017) proposed a WCA hybridisation and dynamic parameter adaptation that aims to improve the WCA’s ability to dynamically adjust its parameters. The results showed that there was a significant improvement in the WCA [64].

Moradi, Sadollah et al. (2017) introduced a metaheuristic optimiser that they called the MOWCA to study the effective frontiers that are related to the standard mean–variance portfolio optimisation model. The evaluation of the performance was done in comparison with the multi-objective PSO (MOPSO). According to the results, MOWCA proved to be an effective optimiser that can be used for portfolio optimisation problems [65].

Pahnehkolaei, Alfi et al. (2017) proposed a gradient-based WCA (GWCA) with evaporation rate in order to enhance the basic WCA. This was achieved by combining it with a local optimiser. Based on the results, the proposed solution performed better compared with other optimisers. The experimental results demonstrated the efficacy of the proposed model [66].

WCA metaheuristic can be utilised in solving optimisation problems that need short computation time and a high degree of accuracy. Researchers were able to develop WCA as an optimiser in numerous domains. WCA and other metaheuristic techniques have been compared. WCA exhibited good results, which resulted into the improvement of the WCA algorithm in solving various optimisation problems.

3 Recommended methods: water cycle algorithm with probabilistic neural network

In this paper, the authors recommend the usage of the WCA evolutionary algorithm to enhance the standard PNN classifier’s performance. Thus, a new hybrid algorithm, the WCAPNN, is recommended for addressing the classification issues.

Figure 1 depicts the structure of a PNN Classifier. Training dataset had been utilised for training the PNN. Testing datasets were then utilised for classifying the unclassified instances. The classification precision was ultimately computed as per Eq. (16).

Fig. 1
figure 1

Structure of probabilistic neural network classifier

As previously mentioned, Eskandar, Sadollah et al. (2012) introduced the WCA metaheuristic. It begins with a preliminary population that is referred to as raindrops. Forming the variables into an array defined as Raindrop = \(\left[{x}_{1},{x}_{2},{x}_{3},\ldots ,{x}_{n}\right]\)is essential for every solution [52].

A sea is formed by the best raindrops, and the rest form rivers and streams. Each river or sea absorbs raindrops coming from the streams based on their magnitude. Moreover, rivers tend to move to the sea. WCA’s main advantage is that it functions in every direction [62]. This advantage can be used to enhance the NN performance by increasing the convergence speed and lessening the trapped local minimum chance.

3.1 Steps and flowchart of the water cycle algorithm

WCA can be implemented using the following 12 steps [56]:

Step 1 initiation of the input parameters: \({N}_{sr},{N}_{pop},{d}_{max}, \mathrm{m}\mathrm{a}\mathrm{x}\_iteration\).

Step 2 random initiation of the population, streams (raindrops), rivers, and sea with the help of the following equations:

$$\begin{aligned} Population \; of \;raindrops &=\left[\begin{array}{c} {sea}_{1}\\ {river}_{2}\\ {Raindrop}_{3}\\ .\\ .\\ {Raindrop}_{Npop}\end{array}\right]\hfill \\ &=\left[ {\begin{array}{cccc} {\mathop x\nolimits_{1}^{1} } & {\mathop x\nolimits_{2}^{1} } & \cdots & {\mathop x\nolimits_{{N{\text{var}} }}^{1} } \\ {\mathop x\nolimits_{1}^{2} } & {\mathop x\nolimits_{2}^{2} } & \ldots & {\mathop x\nolimits_{{N{\text{var}} }}^{2} } \\ \vdots & \vdots & \vdots & \vdots \\ {\mathop x\nolimits_{1}^{Npop} } & {\mathop x\nolimits_{2}^{Npop} } & \cdots & {\mathop x\nolimits_{{N{\text{var}} }}^{Npop} } \\ \end{array} } \right].\end{aligned}$$
(1)

Step 3 assess the cost value for every raindrop by the following cost function (C) stated as:

$${C}_{i}={Cost}_{i}=f\left({x}_{1}^{i},{x}_{2}^{i},\ldots ,{x}_{Nvar}^{i}\right) i=\mathrm{1,2},3,\ldots ,{N}_{pop}$$
(2)

where \({N}_{pop}\) and \({N}_{vars}\) are the initial population (number of raindrops) and the number of design variables, respectively. Every decision variable values (\({x}_{1}, {x}_{2}, {x}_{3},\ldots , {x}_{Nvar}\)) is presented as a floating point number. The best raindrop is chosen for the sea and rivers. \({N}_{sr}\) is the summation of the number of raindrops in the sea and rivers:

$${N}_{sr}=Number of Rivers+1$$
(3)

The remaining raindrops are calculated by the following equation:

$${N}_{Raindrops}= {N}_{pop}-{N}_{sr}$$
(4)

Step 4 define the intensity of the rivers and sea by the equation below:

$${NS}_{n}=round \left\{\left|\frac{{Cost}_{n}}{{\sum }_{i=1}^{Nsr}{Cost}_{i}}\right|\times {N}_{Raindrops}\right\}, n=\mathrm{1,2},\ldots,{N}_{sr}$$
(5)
$${C}_{n}={Cost}_{n}-{Cost}_{Nsr+1}, n=\mathrm{1,2},\ldots ,{N}_{sr}$$
(6)

where \({NS}_{n}\) is the number of streams which flow into the sea or rivers.

Step 5 the streams flow into the rivers as per the following equation:

$${X}_{Stream}\left(t+1\right)={X}_{Stream}\left(t\right)+rand\times C\times \left({X}_{sea}\left(t\right)-{X}_{Stream}\left(t\right)\right)$$
(7)
$${X}_{Stream}\left(t+1\right)={X}_{Stream}\left(t\right)+rand\times C\times \left({X}_{River}\left(t\right)-{X}_{Stream}\left(t\right)\right)$$
(8)

where \(rand\) typifies a uniformly distributed random value within the range of \(\left[\mathrm{0,1}\right]\) and \(C\) is a value near to 2. \(\mathrm{T}\mathrm{h}\mathrm{e} C\) value is larger than 1 so as to stop the streams from moving in multiple directions.

Step 6 the rivers move towards the sea as per the following equation:

$${X}_{River}\left(t+1\right)={X}_{River}\left(t\right)+rand\times C\times \left({X}_{Sea}\left(t\right)-{X}_{River}\left(t\right)\right)$$
(9)

Step 7 exchange the position of a river with a stream which provides the best solution.

Step 8 if a river detects a better solution than the sea, the position of the river is traded with that of the sea (similar to Step 7).

Step 9 check the evaporation condition:

$$\begin{array}{c}if \left|{X}_{Sea}-{X}_{River}^{i}\right|<{d}_{max} or rand<0.1 i=\mathrm{1,2},3,\ldots,{N}_{sr}-1\\ Perform Raining Process\\ end if\end{array}$$
(10)

where \({d}_{max}\) is a very small number to enhance the WCA’s ability.

The next equation is for the streams that move towards the sea to determine near-optimum solutions.

$$\begin{array}{c}if \left|{X}_{Sea}-{X}_{Stream}^{i}\right|<{d}_{max} i=\mathrm{1,2},3,\ldots,{NS}_{n}\\ Perform Raing Process\\ end if\end{array}$$
(11)

The rainfall process is executed for creating a new population in a different direction.

Step 10 check the evaporation condition; in case it is satisfied, rainfall will occur to generate a new stream:

$${X}_{stream}^{new}\left(t+1\right)=LB+rand\times \left(UB-LB\right)$$
(12)

where the upper bound (UB) and lower bound (LB) depict the upper and lower boundaries, respectively.

Step 11 reduce the value of \({d}_{max}\) which is a user-defined parameter by using the equation below:

$${d}_{max}\left(t+1\right)={d}_{max}\left(t\right)-\frac{{d}_{max}\left(t\right)}{Max.iteration}$$
(13)

Step 12 check the convergence criterion. In case it is satisfied, the algorithm will have stopped; else, return to Step 5.

This study used the WCA to find the optimal weights that can be used with the PNN algorithm. Hence, it proposed a new hybrid algorithm called the WCA–PNN to address the classification problems. Figure 2 demonstrates that the process starts by the random generation of the initial weights by the PNN algorithm. The input values are then multiplied by the corresponding weights \({w}_{(ij)}\), based on the values determined by the PNN model.

Fig. 2
figure 2

Representation of initial weights

Figure 3 shows the proposed WCA–PNN’s structure. It is made up of two parts: the first part involves the PNN where the training data is utilised. The tested data then undergoes categorisation. Equation (16) is used to compute the accuracy. Then, the PNN weights are fine-tuned using the WCA. It will then calculate the accuracy of the newly classified data. This process repeats until it meets the termination criterion.

Fig. 3
figure 3

Flowchart of proposed WCA–PNN

The proposed hybrid algorithm has taken into account the exploitation (intensification) and exploration (diversification) in order to achieve high accuracy. During the WCA’s exploitation phase, the rivers represent good solutions and the sea contains the best solution. The rivers and the sea serve as leader points that guide the streams away from unfit areas and to better locations. Equations (7)–(9) are used to compute for the locations of new rivers and streams. During the exploration phase, the WCA’s evaporation condition in Eqs. (10) and (11) prevents the early convergence to the local optima and generates a new stream, as demonstrated in Eq. (12).

However, population initialisation is the main principle that has to be dealt with in diversification. WCA uses a greedy heuristic to generate the initial population. The initial solution is also determined based on the cost intensity landscape that is related to the optimisation problem. The most important disadvantage, however, is that the initial population may have a tendency to lose its diversity, which will result into population stagnation and generate a premature convergence.

Enhanced WCA (E-WCA) has also been proposed. Its basis is the same previously described procedure for the WCA. However, the limitation of WCA is addressed by statistically improving the exploration process. E-WCA generates a new population at a far point based on the current minimum. It aims to enhance the diversity prevent population stagnation, and avoid the generation of premature convergence. The ultimate goal is to create a balance between the exploration and exploitation of the search space.

$$\begin{array}{c}if \left|{X}_{Sea}-{X}_{River}^{i}\right|<{d}_{max} or randn<0.9 i=\mathrm{1,2},3,\ldots,{N}_{sr}-1\\ Perform Raining Process\\ end if\end{array}$$
(14)
$${X}_{stream}^{new}\left(t+1\right)=LB+randn\times \left(UB-LB\right)$$
(15)

Thus, the E-WCA is considered to be an extension of the WCA that has the ability to prevent premature convergence by evading the local minima trap for global optimisation.

4 Experiments, results, and discussions

The testing of the proposed algorithms’ performances was conducted through their application to the 11 freely downloadable UCI datasets for binary classification. For each dataset, 10 runs were independently conducted (based on the description from the dataset website). This is demonstrated in Table 1 [67].

Table 1 Characteristics of the datasets

As illustrated in Table 2, primary examinations were conducted for the regulation of the suitable parameters for the WCA algorithm. Thereafter, these values were then utilised during the conduct of the experiments that were described in this paper. Values similar to the FA-PNN algorithm were used [50].

Table 2 Input parameter settings

The proposed hybrid model’s classification quality was assessed by figuring four counts. True positives (TPs) refer to the class with the number of correctly assigned records; true negatives (TNs) refer to the number of correct instances that are not part of the class; false positives (FPs) refer to the number of instances that are incorrectly assigned; false negatives (FNs) refer to the positive tuples that are labelled incorrectly. Table 3 presents these four counts for the binary classification [68].

Table 3 Confusion matrix of a two-class problem

Classification accuracy [shown in Eq. (16)] is a term that refers to a statistical measure that displays how well the classifier is able to correctly identify the objects towards the labelled classes [69]. The error rate [shown in Eq. (17)] is a measure of the objects that are incorrectly recognised. Furthermore, specificity [shown in Eq. (19)] and sensitivity [shown in Eq. (18)] were also calculated.

$$Accuracy = \frac{TP + TN}{{TP + TN + FP + FN}}$$
(16)
$$ErrorRate = 1 - \frac{TP + TN}{{TP + TN + FP + FN}}$$
(17)
$$Sensitivity = \frac{TP}{{TP + FN}}$$
(18)
$$Specificity = \frac{TN}{{TN + FP}}$$
(19)

4.1 Results for quality of classification

Table 4 illustrates the TP, FP, TN, FN, and the classification accuracy of the proposed E-WCA–PNN and WCA–PNN models, as well as the results of the traditional PNN. Based on the table, the PNN model was outperformed by the proposed hybrid algorithms. It is clear that the E-WCA–PNN model possesses the best classification accuracy for all the datasets because of the E-WCA’s high search capability. This is integrated into the PNN so that the optimal weights can be determined and the enhancement of the performance of the original PNN can be observed accordingly. Thus, E-WCA was able to achieve good balance between the exploration and exploitation capabilities in determining the best parameters since numerous solutions are grouped close to the optimal solution.

Table 4 Classification accuracy, sensitivity, specificity and error rate (%) for PNN, WCA and E-WCA with PNN

Sensitivity is a measure of the TPs that have been correctly recognised, while specificity is a measure of the correctly recognised TNs. For instance, the E-WCA–PNN was able to get a sensitivity of 74.2% in the LD dataset. This means that if one is to conduct a diagnostic examination to identify patients who have the disease, it would be able to positively identify 74.2% of the patients with the disease. The E-WCA–PNN demonstrated 100% sensitivity for the Fourclass dataset. Alternatively, the E-WCA–PNN had 81.8% specificity, which means that if one is to conduct a diagnostic examination to identify patients who do not have the disease, it would correctly identify 81.8% of the patients who do not have the disease as negative. The E-WCA–PNN achieved 100% specificity on the AP, Parkinson’s, Heart, and Fourclass datasets. Based on Table 4, the E-WCA–PNN and WCA–PNN exhibited better performance (based on specificity and sensitivity) compared to the PNN algorithm on six datasets.

4.2 Results for convergence speed

The proposed algorithms’ convergence behaviour upon application to the 11 datasets was also examined. For each of the proposed algorithms, a total of 200 iterations were run because it was apparent that the accuracy did not improve after the iteration is 150. For every test dataset, the convergence of the proposed algorithms with respect to the optimal value [classification accuracy (%)] is demonstrated in Fig. 4.

Fig. 4
figure 4figure 4

Convergence characteristics of WCA and E-WCA

Figure 4 illustrates how WCA converges to local optima from the first iteration. This was achieved because WCA is a kind of greedy heuristic algorithm that gathers several solutions during the initialisation of the population before losing its diversity and causing stagnation of the population and premature convergence. It is believed by the authors that better solutions can be achieved by a stable and fast convergence. This is achieved by improving the WCA and enhancing the randomisation so that the diversity of the population can be maintained. Doing so will delay premature convergence and strike a balance between exploitation and exploration, which helps prevent getting stuck at local optima.

4.3 Results for significance test

The Wilcoxon test was used to analyse the algorithms and evaluate the classification accuracy. Table 5 shows the statistics for the WCA and FA accuracy. The study also explored the performance of the WCA and E-WCA to determine if it has any statistical difference from the FA. This was done by conducting a Wilcoxon test on classification accuracy with a significance interval of 95% (α = 0.05). The obtained p-values are presented in Tables 7 and 8.

Table 5 Statistics of WCA and FA

The statistical information that is given in Tables 6 and 7 demonstrates that based on accuracy, majority of the WCA results have a statistical difference (p-value < 0.05) from the results of FA. However, no significant difference was observed for AP and BC datasets.

Table 6 p-values of t-test for accuracy of WCA and FA
Table 7 p-values of t-test for accuracy of E-WCA and FA

Furthermore, Tables 6 and 7 presented statistical information that shows that in terms of accuracy, majority of the WCA and E-WCA results had a statistical difference (p-value < 0.05) from the results of FA. The performance of the WCA was therefore significantly better compared to that of the FA because the p-value of the WCA for most of the datasets is less than 0.0001. The simulation results are also a confirmation that E-WCA is an appropriate method for classification problems because it exhibits good performance since it did not rank last for all the datasets that were tested (based on classification accuracy).

Figure 5 shows the box plots that provide an explanation for the resolution quality distribution obtained by the WCA, FA and E-WCA for the 11 datasets.

Fig. 5
figure 5

Box plots for WCA, E-WCA and FA

4.4 Comparison with methods in the literature

For performance assessment with comparable classification models, this paper utilises the same set of datasets and approaches for evaluating the classification accuracy attained by the recommended hybrid WCA–PNN and E-WCA–PNN. The listed methodologies include: flexible neural-fuzzy inference system [70], BP [71], ANNs [72], cover learning using integer programming 3 (CLIP 3) [73], C4.5 [4], convexity based algorithm (CBA) [74], and hybrid simplex-GA [75], firefly with PNN algorithms [50]. The best outcomes are depicted in bold [76]. The results of this comparison are presented in Table 8.

Table 8 Comparison of results using WCA, E-WCA and other classification models

Table 8 shows that with respect to classification accuracy, E-WCA is ranked first for majority of the tested datasets. Furthermore, WCA and E-WCA were categorised in the Fourclass dataset with 100% accuracy. In general, it was clear that the E-WCA outperformed the other approaches and was able to achieve the best classification accuracy based on all the tested datasets. The experimental results also showed that the E-WCA–PNN performed better than all the 11 tested datasets (based on the classification accuracy).

It is believed by the authors that the E-WCA was able to perfectly diversify the initial solution, which forced the solutions to be nearer to each other. It also provided a balance to the exploitation and exploration processes, which prevented population stagnation and premature convergence. Therefore, it enhanced the performance of PNN by determining the values of the near optimal weights.

5 Conclusion

This paper mainly aimed to suggest a new method that is capable of establishing good-quality solutions that can be used for classification problems. The WCA is a population-based metaheuristic that simulates the natural water cycle taking place in the real world. Therefore, it can optimise the PNN’s weight values. The improved exploitation and exploration capabilities of the WCA give it the ability to accomplish better results compared to some other algorithms when a large search space is being explored. Therefore, it can find efficient and effective solutions that can be used to answer numerous complex classification problems.

The proposed hybrid E-WCA–PNN was then applied to 11 benchmark UCI datasets. During the first iteration, a random parameter was generated by the PNN. Then, the improvement occurred in the subsequent iterations when the PNN weights were optimised by the WCA. For the experimental test, PNN was served as a single classifier and the gathered results were then utilised for the measurement of the classification accuracy, sensitivity, error rate, and specificity. The values obtained were then compared with the results from the proposed hybrid model so that the objectives of this research can be achieved.

It was proven by the experimental results that the proposed models could perform better than the PNN classifier when the latter was solely applied to all the tested datasets. In summary, the obtained results demonstrated that the proposed hybrid algorithms were efficient and effective and they were capable of producing a high-quality classification solution due to the fact that they were able to achieve higher convergence speed and better classification accuracy compared to other comparable methods. As a future study, the authors have planned to make a E-WCA hybrid with other search algorithms that possess high exploration, so that a balance between the exploitation and exploration during optimisation can be achieved and population diversity can be maintained.