iconOpen Access

ARTICLE

crossmark

A Double Adaptive Random Spare Reinforced Sine Cosine Algorithm

Abdelazim G. Hussien1,2, Guoxi Liang3, Huiling Chen4,*, Haiping Lin5,*

1 Department of Computer and Information Science, Linköping University, Linköping, 4189, Sweden
2 Faculty of Science, Fayoum University, Fayoum, 63514, Egypt
3 Department of Information Technology, Wenzhou Polytechnic, Wenzhou, 325035, China
4 Artificial Intelligence, Wenzhou University, Wenzhou, 325035, China
5 College of Information Engineering, Hangzhou Vocational & Technical College, Hangzhou, 310018, China

* Corresponding Authors: Huiling Chen. Email: email; Haiping Lin. Email: email

Computer Modeling in Engineering & Sciences 2023, 136(3), 2267-2289. https://doi.org/10.32604/cmes.2023.024247

Abstract

Many complex optimization problems in the real world can easily fall into local optimality and fail to find the optimal solution, so more new techniques and methods are needed to solve such challenges. Metaheuristic algorithms have received a lot of attention in recent years because of their efficient performance and simple structure. Sine Cosine Algorithm (SCA) is a recent Metaheuristic algorithm that is based on two trigonometric functions Sine & Cosine. However, like all other metaheuristic algorithms, SCA has a slow convergence and may fail in sub-optimal regions. In this study, an enhanced version of SCA named RDSCA is suggested that depends on two techniques: random spare/replacement and double adaptive weight. The first technique is employed in SCA to speed the convergence whereas the second method is used to enhance exploratory searching capabilities. To evaluate RDSCA, 30 functions from CEC 2017 and 4 real-world engineering problems are used. Moreover, a nonparametric test called Wilcoxon signed-rank is carried out at 5% level to evaluate the significance of the obtained results between RDSCA and the other 5 variants of SCA. The results show that RDSCA has competitive results with other metaheuristics algorithms.

Keywords


Supplementary Material

Supplementary Material File

1  Introduction

Recently, an enormous number of meta-heuristics algorithms have been proposed which simulate natural-based phenomena or biological behavior [16]. These meta-heuristics have been successfully applied to various domains of medical science [710], finance [1113], energy [1416], engineering [1721], agriculture [22,23] and education [2426]. Some examples of these algorithms are Artificial Bee Colony (ABC) [27], Genetic Algorithm (GA) [28], Ant Colony Optimization (ACO) [29], Particle Swarm Optimization (PSO) [30], Cuckoo Search (CS) [31,32], Grey Wolf Optimizer [33], Bat Algorithm (BA) [34], Runge Kutta Optimizer (RKN) [35], Colony Predation Algorithm (CPA) [36], Crow Search Algorithm (CSA) [37], Weighted Mean of Vectors [38], Snake Optimizer (SO) [39], Hunger Games Search (HGS) [40], Whale Optimization Algorithm (WOA) [41], Symbiotic Organism Search (SOS) [42], Lightning Search Algorithm (LSA) [43], Moth-Flame Optimization (MFO) [44], Slime Mould Algorithm (SMA) [45], Ant Lion Optimizer (ALO) [46], Harris Hawks Optimization (HHO) [47,48], Remora Optimization Algorithm (ROA) [49], and Wild Horse Optimizer (WHO) [50].

Sine Cosine Algorithm (SCA) is a novel population-based algorithm that has been introduced to solve real-world problems. For example, Elaziz et al. [51] developed the use of SCA to optimize the optimal parameters of Random Vector Functional Chain (RVFC) networks and use the obtained optimal model for predicting ACE inhibitory activity. Li et al. [52] developed a novel SCA-SVR model using the SCA in order to select the penalty & parameters of the kernel in SVR to improve the performance of generalization to unknown data. The model was able to have the optimal values of the SVR parameters with good results. Nayak et al. [53] introduced an enhanced Extreme Learning Machine (ELM) approach for the sine cosine algorithm (MSCA-ELM) and designed a new automated Magnetic Resonance Imaging (MRI) brain pathology detection system. Wang et al. [54] introduced a hybrid wavelet neural network based on the proposed Multi-Objective Sine Cosine Algorithm (MOSCA), which was successfully applied in wind speed prediction. Dasgupta et al. [55] introduced a newly designed SCA for determining the optimal voltage regulator systems of proportional-integral differential controller parameters, which can effectively tune the PID controller parameters. Chen et al. [16] combined opposition-based learning and local search strategies with SCA to estimate photovoltaic models parameters. Also, Sahlol et al. [56] tried to improve fish liver Enzymes prediction by using SCA to train feedforward neural networks. Also, Sindhu et al. [57] used an elitism strategy with SCA to solve the feature selection problem. Likewise, SCA has been applied to other problems e.g., predicting cervical hyperextension injury [58], predicting master students’ intention [26], hydrothermal scheduling [59], and image segmentation [60].

Many variants of SCA have been proposed such as a multi-objective SCA (MO-SCA) which was proposed by Tawhid and Savsani [61]. Also, Issa et al. [62] tried to design an adaptive SCA version that is combined with PSO called (ASCA-PSO). Likewise, Sine Cosine Crow Search Algorithm (SCSCA) [63], hybrid algorithms between sine cosine and particle swarm algorithm known as (H-PSO-SCAC) [64], hybrid algorithm between SCA and differential evolution (SCA-DE) [65], Opposition-Based SCA (OBSCA) [66], enhanced sine cosine algorithm called (MSCA) [67], and orthogonal-learning driven multi-swarm sine cosine called (OMGSCA) [68]. Abd Elaziz et al. [68] presented an improved version of SCA that incorporated backward learning to increase the exploration search space to generate more accurate solutions. Guo et al. [69] presented a Riesz derivative fractional sinusoidal algorithm abbreviated as that is based on Riesz fractional derivative variation mechanism and applyed it to the design of welded beams and pressure vessels. Gupta et al. [70] developed a novel and improved sine cosine algorithm using cross-exploitation techniques with the integration of individual optimal states of individual solutions, global search, and self-learning mechanisms and applied it effectively to multi-stage threshold segmentation in images. Gupta et al. [71] developed an improved SCA that is based on the logarithmic and adaptive components of the perturbation rate. Their algorithm was employed in order to solve five different engineering optimization problems and effectively solve the global optimization problem. Issa et al. [64] introduced a modified SCA version in which SCA was combined with PSO. This algorithm was used to solve the pairwise local alignment problem with the aim of looking for the longest continuous substring between two biological sequences, which has shown good performance in accuracy and calculation time. Long et al. [72] introduced inertial weights and a novel SCA version known as (ISCA) and applied it in solving high dimensional problems based on a Gaussian function with a nonlinear transformation parameter decreasing strategy. Compared to traditional SCA and other population-based algorithms, the ISCA algorithm has faster convergence and better ability to escape from local optimal solutions. Nenavath et al. [67] proposed a new differential evolutionary sine and cosine mixture optimization algorithm and confirmed its accuracy under various challenging conditions.

Although the success of all mentioned variants of SCA, it still has low convergence and may struggle in local optimal regions. This motivated us to propose a novel variant of SCA called RDSCA that employed two strategies: 1) double adaptive weight strategy in order to achieve a good balance between exploitation and exploration by using two weights w1 in the first half of iterations and w2 in the last half of iterations. 2) random spare strategy in which each individual in search space can replace the optimal solution in each dimension which helps in escaping from local optima. Moreover, 30 functions from CEC 2017 and 4 real-world constrained engineering problems are used to validate our algorithm. Also, a comparison with many SCA variants has been done with RDSCA. Chen et al. [73] has used the same strategies with WO and they argued that it has a very good results. So, we test the same mechansim on SCA.

To the best of our information & knowledge and according to the literature, this is the first time that these two strategies were introduced into SCA. The main contributions of this paper are shown below:

•   A new metaheuristic algorithm, called RDSCA, is proposed based on a random replacement and double adaptive weight strategy, combined with SCA.

•   In classical CEC2017 benchmark functions, experimental results with SCADE [74], OBSCA [68], CGSCA [75], ASCA_PSO [64] and CESCA [26] demonstrate that the RDSCA algorithm has efficient performance and competitive ability in optimization search.

•   To verify the performance of the algorithm on real-world problems, RDSCA’s experimental results on four engineering problems demonstrate that the optimizer can effectively solve complex real-world problems.

The rest of this paper is organized as follows: the basic SCA and its mathematical equation are given in Section 2 whereas RDSCA is discussed comprehensively in Section 3. Section 4 shows both experiments & the results whereas the conclusion & possible future works are given in Section 5.

2  Sine Cosine Algorithm

Recently many new meta-heuristics algorithms have been developed and have shown great potential in tackling complex optimization problems such as image segmentation [76,77], PID optimization control [78], economic emission dispatch problem [79], expensive optimization problems [80], combination optimization problems [81], bankruptcy prediction [12,82], plant disease recognition [83], traveling salesman problem [84], feature selection [85,86], multi-objective problem [87], detection of foreign fiber in cotton [88,89], medical diagnosis [7,8], fault diagnosis [90,91], parameter optimization for machine learning models [92,93], prediction problems in educational field [24,25], constrained optimization problems [94,95], scheduling problem [96,97], and object tracking [98,99]. SCA can be considered as one of the most powerful and well-known meta-heuristics algorithms which was firstly proposed in 2016 [100]. The main idea in SCA is Sine & Cosine operator, which update agent movement according to Eqs. (1a), (1b):

xit+1=xit+r1×sin(r2)×|r3pitxit| if r4<0.5(1a)

xit+1=xit+r1×cos(r2)×|r3pitxit| if r40.5(1b)

where r2, r3, and r4 are random numbers in the interval (0,1], pit is the best solution position obtained in the i-th iteration, and r1 can be obtained from the following equation:

r1=ataT(2)

where a refers to a constant number, and T, t refer to the maximum iteration number and the current iteration, respectively. Also, flowchart of SCA is given in Fig. 1.

images

Figure 1: Flowchart of SCA

3  Proposed Algorithm

3.1 Architecture of RDSCA

In this section, the structure of the suggested algorithm called RDSCA has been illustrated in detail. RDSCA is a modification of SCA with 2 strategies: 1) random spare/replacement. 2) double adaptive weight. The pseudo-code and flow chart are given in Algorithm 1 and Fig. 2. The integration of two mentioned strategies has a great impact on SCA in search capabilities and convergence curve.

images

:

images

Figure 2: Flowchart of RDSCA

3.1.1 Random Spare/Replacement

In this strategy, the current individual with n-th dimension position will be replaced with the corresponding dimension of the optimal individual vector. A certain probability is used to implement a random spare strategy, since not any dimension value in the optimal individual is optimal. Cauchy random is used as a ratio between the current evaluation number to total evaluation number.

3.1.2 Double Adaptive Weight

In this strategy, similar to PSO [30], a double adaptive weight was used to improve the ability of local & global search w1, w2 can be calculated from Eqs. (3) and (4). Furthermore, Fig. 1 in the supplement analyzes the balance of RDSCA and SCA.

w1=(1FEsMaxFES)1tan(π×(rand0.5)×sMaxFEs)(3)

w2=(22FEsMaxFES)1tan(π×(rand0.5)×sMaxFEs)(4)

where FEs is the current number of iterations, MaxFES is the maximum number of iterations of the algorithm, tan(π×(rand0.5) denotes the Cauchy random number, s changes with the local optimum, and s automatically increases by 1 when an individual is not updated, otherwise the value of s is subtracted by half when an individual is updated. w1 and w2 are not linearly decreasing, but fluctuate in frequency as the algorithm falls into the local optimum, and the fluctuation range of w1 is [0, 1], w2 takes values in the range [0.5, 1]. Therefore, w1 is mainly used in the first half of the algorithm iteration to improve the global search ability of the algorithm, and w2 is used in the second half of the algorithm iteration to improve the ability of the algorithm to jump out of the local optimum. Eq. (5) is used in the first half of iterations whereas, Eq. (6) is used in the last half of iterations.

xit+1={xit+w1×r1×sin(r2)×|r3pitxit| if r4<0.5xit+w1×r1×cos(r2)×|r3pitxit| if r40.5(5)

xit+1={xit+w2×r1×sin(r2)×|r3pitxit| if r4<0.5xit+w2×r1×cos(r2)×|r3pitxit| if r40.5(6)

3.2 The Proposed RDSCA

In the algorithmic optimization search process, certain dimensional vectors of the current individual may be responsible for the superior quality of the individual, and in contrast, certain dimensions may also lead to the poor quality of the individual. This is explained by the fact that the result of the mapping from the decision space to the objective space is determined jointly by multiple dimensions of the decision variables. Therefore, we go to adopt the values of some dimensions of the current optimal individual to replace the values of the corresponding dimensions of the current individual with a certain probability in the optimization process of the algorithm. In addition, the timing of random replacement is also very important, and we use the mathematical description in lines 10 to 13 of Algorithm 1 to define a random replacement probability. The mathematical description is mainly used to determine whether the individual adopts the random replacement strategy through the Cauchy random number and the current number of iterations. At the beginning of the algorithm iteration, the Cauchy random number can improve the randomness, so we have a higher probability to replace some dimensions of the individual, thus speeding up the convergence of the algorithm; at the later stage of the algorithm iteration, the mathematical description in the third line of the pseudo-code will have a smaller probability to replace the dimensions of the individual, resulting in Individuals have more vectors of random dimensions, thus reducing the possibility of the algorithm falling into a local optimum.

In addition, the proposed RDSCA algorithm in which we introduce an double adaptive weight strategy is shown in lines 16 to 22 of Algorithm 1. w1 and w2 are affected by the change of individuals into local optimum, which can increase the ability of the algorithm to jump out of local optimum when individuals are detected to be in local optimum. The specific principle is that when an individual is not updated for a long time, the individual is more likely to fall into a local optimum, and then the values of weights w1 and w2 increase, increasing the range of random fluctuations in Eqs. (1a), (1b) and enhancing the ability of the algorithm to jump out of the local optimum. As a result, the algorithm can jump out of the local optimum faster and increase the speed of convergence. The random replacement strategy and the double adaptive weight strategy cooperate with each other in the iterative process of the algorithm, which well balances the algorithm between search and exploitation to improve the convergence ability and the ability to jump out of the local optimum.

In addition, it should be especially noted that RDSCA and SCA use the same algorithm framework, and SCA mainly accomplishes the balance of algorithm exploration and exploitation through Eqs. (1a), (1b), while RDSCA mainly balances algorithm exploration and exploitation through Eqs. (5) and (6). Eqs. (5) and (6) expand the perturbation frequency and amplitude of RDSCA by adding adaptive dual weights w1 and w2 without changing the conversion mechanism of the algorithm from exploration to exploitation.

3.3 Time Complexity of RDSCA

The complexity of any meta-heuristic algorithm is very crucial as it computes its needed time. The time complexity of RDSCA is is depended on the number of individuals (N), maximum iteration number (T), and number of dimensions (Dim). Therefore, RDSCA complexity is as follows:

O(RDSCA)=O(Initialization)+O(Evaluating )+O(Updating individual)=O(N×D)+O(N×D×T)+O(N×D×T)=O(N×D×T)

4  Experiment

In order to verify the effectiveness of the algorithm, a fair experiment is necessary, and the whole process must be controlled and have a verifiable data set [101]. In this section, a discussion on the performance of RDSCA is introduced based on some experiments using CEC 2017, they have been internationally recognized test data in the field of metaheuristic algorithms. References [102,103] are a benchmark and a number of compared algorithms. All experimental results are the average of the results of 30 independent runs of the algorithm on each benchmark function. All experiments of the algorithms were coded and run on the MAT-LAB R2014b software platform.

4.1 Benchmark Test Functions

In order to evaluate the performance of the proposed RDSCA, 30 functions from CEC 2017 benchmark were used. These functions have really complex parameters and a variety of real optimization difficulties that are close to reality. They can be classified into unimodal (F1–F3), multi-modal (F4–F10), hybrid (F11–F20), and composite (F21–F30) as given in Table S1 in the supplement, where ‘’Types” denotes the type of function; unimodal indicates that there is only one global optimum solution and no local optimum, which can test the convergence speed of the algorithm; multi-modal contains multiple local optima with one and only one global optimum; hybrid and composite functions are more complex, which often have multi-modal nature and have more difficult to jump out of the local optimum, which requires a better balance between exploration and exploitation of the algorithm. “Name” denotes the name of the function; “Opt” denotes the optimal objective value of the function. In all experiments, the number of population (n) is set to 30, dimension numbers (dim) is equal to 30, and the number of maximum evaluation of fitness function (MaxFEs) has been set to dim * 10000 = 300000 as it is shown in Table 1.

images

4.2 Comparative Algorithms

To evaluate the performance of RDSCA, a comparison against many other variants of SCA has been done using 30 functions from CEC 2017. Modified SCA versions are chosen for comparison purpose, which are mentioned as bellow:

•   SCADE [74]: differential evolution (DE) operators have been employed in SCA as local search operators.

•   OBSCA [68]: opposition-based learning (OBL) is used in the initialization and updating phases.

•   CGSCA [75]: exponential decreasing conversion parameter and linear decreasing inertia weight methods are employed within SCA to achieve a good balance between exploration and exploitation. Moreover, a greedy Levy mutation technique is also applied.

•   ASCA_PSO [62]: ASCA-PSO is an algorithm that hybridizes PSO with SCA in order to have more exploitation properties.

•   CESCA [26]: chaos theory concepts have been employed in SCA.

4.3 Measures of the Performance

Here, a number of statistical metrics are used to compute the RDSCA performance.

•   Average (mean):

Avg=1Ni=1Ngi(7)

•   Std:

SD=1N1i=1N(giAvg)2(8)

where N refers the number of total evaluated times, gi is the value of the fitness function at i-th iteration.

•   Min: the best solution obtained over all running.

best=mini=1Ngi(9)

where M refers to the iteration number to perform the experiments and gi refers to the result obtained from the fitness function at i-th iteration and Avg refers to the average of all running.

•   Max: the worst solution obtained over all running.

Worst=maxi=1Ngi(10)

where N refers to iteration number to perform the experiments and gi is the result obtain fron the fitness function at i-th iteration.

4.4 Discussion

4.4.1 Comparison between RDSCA and Other Algorithms

The comparison results between RDSCA and other SCA variants mentioned in Section 4.2 are shown in Table 2. From this table, it is notable that the suggested algorithm achieved better fitness value. RDSCA achieved the best average in 24 functions from 30 functions, second best in 4 other functions (F2, F11, F18, and F24) and the third best in F14. Furthermore, RDSCA ranked first in term of standard deviation in 19 functions and the second best in the other 11 functions.

images

Fig. 3 shows the convergence curve for some functions from all types (unimodal, multimodal, hybrid, and composite). It can be observed from this figure that RDSCA has a better speed convergence compared to other algorithms.

images

Figure 3: The average of convergence curves of unimodal & multi-modal functions

Also, Fig. 4 shows the 2-D appearance of the function in the first column whereas the search history is shown in the second one where we can observe that RDSCA’s search agent tends to exploit the promising regions of search space more extensively. The third column shows the first particle trajectory where we can see the first agent changes in the first dimension and the fourth and fifth columns show RDSCA’s average fitness and convergence curve, respectively. Furthermore, the result of banlance analysis of RDSCA and SCA is shown in Figure S1 in the supplement, which also shows the superiority of RDSCA.

imagesimages

Figure 4: (a) The average of convergence curves of unimodal & multi-modal functions 2; (b) The average of convergence curves of unimodal & multi-modal functions 2

4.4.2 Statistical Analysis

Besides the performance measured mentioned in Section 4.3, a non-parametric test called Wilcoxon signed-rank (WSR) [104] is used to be able to prove the significant superiority of the RDSCA. WRS considers all results as one group and then ranks are assigned to all scores in each group then, WRS test sums each group rank.

All algorithms are tested in all test functions for WSR. Each algorithm is performed 30 times independently on each test function, and the resulting solutions are available as data samples. WSR can test whether the two algorithms are significantly different on that test function based on these samples. First, we start with the null hypothesis, which is the assumption that there is no significant difference in the performance of the two algorithms. When p<0.05, the null hypothesis is rejected, and when p>0.05, the null hypothesis is accepted. Therefore, the two algorithms are significantly different when the p-value is greater than 0.05. Table S2 in the supplement represents the p-values for RDSCA and each comparison algorithm on each test function. From the table, we can find that most of the p-values are less than 0.05, therefore, we can conclude that RDSCA and other comparison algorithms are significantly different on the CEC2017 test functions and that RDSCA outperforms the comparison algorithms from the results in Table 2.

4.5 Engineering Problems

To validate our proposed meta-heuristic which known as RDSCA, we apply it to solve real engineering constrained problems which have various inequality and equality constraints. Furthermore, best values of such parameter of these problems remain unknown. Here, four different engineering optimization problems are used in order to validate RDSCA namely: tension/compression spring design (TCSD), welded beam design (WBD), pressure vessel design (PVD), and I-beam design (IBD).

4.5.1 TCSD Problem

The first engineering problem is called TCSD [105] (see Appendix A for the mathematical formulation). This problem aims to find the optimal and minimum weight and includes 3 design variables: mean coil diameter (D), active coils number (N), and diameter of wire (d).

Table S3 in the supplement shows the results of RDSCA against other peers. It can be noticed that the proposed algorithm achieved the best results.

4.5.2 WBD Problem

WBD problem is a well-known metal problem [106] which aims to obtain welded beam that has the minimum cost of fabrication. The mathematical modeling of WBD can be found in Appendix B. Here a comparison of our proposed approach has been done against other peers as shown in Table S4 in the supplement. It is obviously noted from the above mentioned table that RDSCA has got the minimum optimal solution, i.e., minimum fabrication cost.

4.5.3 PVD Problem

The purpose of this unconstrained engineering problem is to find the minimum cost of the cylindrical pressure vessel [107]. The PVD has 4 variables: head thickness (Th), shell thickness (Ts), inner radius (R), and cylindrical section length (L). Appendix C shows mathematical equations of PVD.

Table S5 in the supplement shows the comparison results among RDSCA, other peers. It is obviously noticed that RDSCA ranked first.

4.5.4 IBD Problem

In this section, our proposed algorithm is used to find the optimal design of I-beam problem [108] in order to have minimum deflection. This problem has four design parameters: height, length, and 2 thickness. The mathematical formulation is founded in section Appendix D. The results of RDSCA against other algorithms are shown in Table S6 in the supplement where we can observe that RDSCA ranked first.

5  Conclusion & Possible Future Works

In this study, a new metaheuristic algorithm, called RDSCA, is proposed by adding double adaptive weight and random replacement strategy to the algorithmic framework of SCA, where random replacement are used in the process of iterative optimization of the algorithm, and each individual has a certain probability to select a dimensional value and the corresponding dimension of the current optimal value for replacement. The basic principle of this strategy is that the cause of poor individual fitness values may be the value of a single dimension of an individual. And this method can effectively increase the convergence speed of the algorithm because the individuals are using as much information as possible from the current optimal solution. The double adaptive weight strategy mainly uses two adaptive weights w1 and w2 to enhance the ability of the algorithm to jump out of the local optimum, where the updates of w1 and w2 are related to the frequency of individuals falling into the local optimum and the Corsi random number. Once an individual falls into a local optimum, the algorithm can jump out faster, thus improving the convergence speed. Based on this, we experimentally compare RDSCA with other mainstream metaheuristic algorithms on CEC2017 and four restricted real-world engineering problems, and the experimental results prove that RDSCA has stronger performance and strong competitiveness.

However, RDSCA and SCA’s use the same algorithm framework, which also means that the algorithm is explored and developed in the same way of conversion as the original SCA, which may not be able to better balance the exploration and development of the algorithm when optimizing complex composite functions. In the future, we will use new techniques to improve the ability of algorithms to balance exploration and development. In the future, we will apply this version to many applications in different fields such as community service [109], multicore systems [110], etc. We can also try to propose a multi-objective version and binary version of RDSCA for other interesting scenarios. Also, it is possible in future to apply these combination of operators to other metaheurstics and comparing them.

Funding Statement: This study was supported in part by the Hangzhou Science and Technology Development Plan Project (Grant No. 20191203B30).

Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study.

References

  1. Feng, Y., Wang, G., Deb, S., Lu, M., & Zhao, X. (2019). Monarch butterfly optimization. Neural Computing and Applications, 31(7), 1995-2014. [Google Scholar] [CrossRef]
  2. Hussien, A. G., Hassanien, A. E., Houssein, E. H., Bhattacharyya, S., Amin, M. (2019). S-shaped binary whale optimization algorithm for feature selection. In: Recent trends in signal and image processing, pp. 79–87. Singapore: Springer.
  3. Wang, G., Coelho, L., Gao, X. Z., & Deb, S. (2016). A new metaheuristic optimisation algorithm motivated by elephant herding behaviour. International Journal of Bio-Inspired Computation, 8(6), 394-409. [Google Scholar] [CrossRef]
  4. Wang, G. G. (2016). Moth search algorithm: A bio-inspired metaheuristic algorithm for global optimization problems. Memetic Computing, 10, 151-164. [Google Scholar] [CrossRef]
  5. Hussien, A. G., Houssein, E. H., Hassanien, A. E. (2017). A binary whale optimization algorithm with hyperbolic tangent fitness function for feature selection. 2017 Eighth International Conference on Intelligent Computing and Information Systems (ICICIS), pp. 166–172. Cairo, Egypt, IEEE.
  6. Hussien, A. G., Hassanien, A. E., Houssein, E. H. (2017). Swarming behaviour of salps algorithm for predicting chemical compound activities. 2017 Eighth International Conference on Intelligent Computing and Information Systems (ICICIS), pp. 315–320. Cairo, Egypt, IEEE.
  7. Xia, J., Wang, Z., Yang, D., Li, R., & Liang, H. (2022). Performance optimization of support vector machine with oppositional grasshopper optimization for acute appendicitis diagnosis. Computers in Biology and Medicine, 143, 105206. [Google Scholar] [PubMed]
  8. Xia, J., Yang, D., Zhou, H., Chen, Y., & Zhang, H. (2022). Evolving kernel extreme learning machine for medical diagnosis via a disperse foraging sine cosine algorithm. Computers in Biology and Medicine, 141, 105137. [Google Scholar] [PubMed] [CrossRef]
  9. Fathi, H., AlSalman, H., Gumaei, A., Manhrawy, I. I., & Hussien, A. G. (2021). An efficient cancer classification model using microarray and high-dimensional data. Computational Intelligence and Neuroscience, 2021, [Google Scholar] [PubMed] [CrossRef]
  10. Hu, J., Zhengyuan, H., Heidari, A. A., Shou, Y., & Ye, H. (2022). Detection of COVID-19 severity using blood gas analysis parameters and harris hawks optimized extreme learning machine. Computers in Biology and Medicine, 142, 105166. [Google Scholar] [PubMed] [CrossRef]
  11. Yu, C., Chen, M., Cheng, K., Zhao, X., & Ma, C. (2022). SGOA: Annealing-behaved grasshopper optimizer for global tasks. Engineering with Computers, 38(5), 3761-3788. [Google Scholar]
  12. Yza, B., Rl, A., Aahc, D., Xin, W. E., & Ying, C. F. (2021). Towards augmented kernel extreme learning models for bankruptcy prediction: Algorithmic behavior and comprehensive analysis-sciencedirect. Neurocomputing, 430, 185-212. [Google Scholar] [CrossRef]
  13. Wang, M., Chen, H., Li, H., Cai, Z., & Zhao, X. (2017). Grey wolf optimization evolving kernel extreme learning machine: Application to bankruptcy prediction. Engineering Applications of Artificial Intelligence, 63, 54-68. [Google Scholar] [CrossRef]
  14. Jiao, S., Chong, G., Huang, C., Hu, H., & Wang, M. (2020). Orthogonally adapted harris hawk optimization for parameter estimation of photovoltaic models. Energy, 203, 117804. [Google Scholar] [CrossRef]
  15. Zhang, H., Heidari, A. A., Wang, M., Zhang, L., & Chen, H. (2020). Orthogonal nelder-mead moth flame method for parameters identification of photovoltaic modules. Energy Conversion and Management, 211, 112764. [Google Scholar] [CrossRef]
  16. Chen, H., Jiao, S., Heidari, A. A., Wang, M., & Chen, X. (2019). An opposition-based sine cosine approach with local search for parameter estimation of photovoltaic models. Energy Conversion and Management, 195, 927-942. [Google Scholar] [CrossRef]
  17. Zhang, H., Liu, T., Ye, X., Heidari, A. A., & Liang, G. (2022). Differential evolution-assisted salp swarm algorithm with chaotic structure for real-world problems. Engineering with Computers, 38, 1. [Google Scholar] [CrossRef]
  18. Shan, W., Qiao, Z., Heidari, A. A., Chen, H., & Turabieh, H. (2021). Double adaptive weights for stabilization of moth flame optimizer: Balance analysis, engineering cases, and medical diagnosis. Knowledge-Based Systems, 214, 106728. [Google Scholar] [CrossRef]
  19. Tu, J., Chen, H., Liu, J., Heidari, A. A., & Zhang, X. (2021). Evolutionary biogeography-based whale optimization methods with communication structure: Towards measuring the balance. Knowledge-Based Systems, 212, 106642. [Google Scholar] [CrossRef]
  20. Hussien, A. G., Oliva, D., Houssein, E. H., Juan, A. A., & Yu, X. (2020). Binary whale optimization algorithm for dimensionality reduction. Mathematics, 8(10), 1821. [Google Scholar] [CrossRef]
  21. Hussien, A. G. (2021). An enhanced opposition-based salp swarm algorithm for global optimization and engineering problems. Journal of Ambient Intelligence and Humanized Computing, 13, 129-150. [Google Scholar] [CrossRef]
  22. Yu, H., Song, J., Chen, C., Heidari, A. A., & Liu, J. (2022). Image segmentation of leaf spot diseases on maize using multi-stage Cauchy-enabled grey wolf algorithm. Engineering Applications of Artificial Intelligence, 109, 104653. [Google Scholar] [CrossRef]
  23. Chen, C., Wang, X., Heidari, A. A., Yu, H., & Chen, H. (2021). Multi-threshold image segmentation of maize diseases based on elite comprehensive particle swarm optimization and otsu. Frontiers in Plant-Science, 12, 789911-789911. [Google Scholar] [PubMed] [CrossRef]
  24. Wei, Y., Lv, H., Chen, M., Wang, M., & Heidari, A. A. (2020). Predicting entrepreneurial intention of students: An extreme learning machine with Gaussian barebone harris hawks optimizer. IEEE Access, 8, 76841-76855. [Google Scholar] [CrossRef]
  25. Zhu, W., Ma, C., Zhao, X., Wang, M., & Heidari, A. A. (2020). Evaluation of sino foreign cooperative education project using orthogonal sine cosine optimized kernel extreme learning machine. IEEE Access, 8, 61107-61123. [Google Scholar] [CrossRef]
  26. Lin, A., Wu, Q., Heidari, A. A., Xu, Y., & Chen, H. (2019). Predicting intentions of students for master programs using a chaos-induced sine cosine-based fuzzy k-nearest neighbor classifier. IEEE Access, 7, 67235-67248. [Google Scholar] [CrossRef]
  27. Karaboga, D., & Basturk, B. (2007). A powerful and efficient algorithm for numerical function optimization: Artificial bee colony (ABC) algorithm. Journal of Global Optimization, 39(3), 459-471. [Google Scholar] [CrossRef]
  28. Holland, J. H. (1992). Genetic algorithms. Scientific American, 267(1), 66-73. [Google Scholar] [CrossRef]
  29. Dorigo, M., Birattari, M., & Stutzle, T. (2006). Ant colony optimization. IEEE Computational Intelligence Magazine, 1(4), 28-39. [Google Scholar] [CrossRef]
  30. Kennedy, J., Eberhart, R. (1995). Particle swarm optimization. Proceedings of ICNN’95–International Conference on Neural Networks, Perth, WA, Australia.
  31. Yang, X. S., Deb, S. (2009). Cuckoo search via lévy flights. 2009 World Congress on Nature & Biologically Inspired Computing (NaBIC), Coimbatore, India.
  32. Abualigah, L., Gandomi, A. H., Elaziz, M. A., Hussien, A. G., & Khasawneh, A. M. (2020). Nature-inspired optimization algorithms for text document clustering a comprehensive analysis. Algorithms, 13(12), 345. [Google Scholar] [CrossRef]
  33. Mirjalili, S., Mirjalili, S. M., & Lewis, A. (2014). Grey wolf optimizer. Advances in Engineering Software, 69, 46-61. [Google Scholar] [CrossRef]
  34. Yang, X. S., & Gandomi, A. H. (2012). Bat algorithm: A novel approach for global engineering optimization. Engineering Computations, 29, 464-483. [Google Scholar] [CrossRef]
  35. Ahmadianfar, I., Heidari, A. A., Gandomi, A. H., Chu, X., & Chen, H. (2021). Run beyond the metaphor: An efficient optimization algorithm based on runge kutta method. Expert Systems with Applications, 181, 115079. [Google Scholar] [CrossRef]
  36. Tu, J., Chen, H., Wang, M., & Gandomi, A. H. (2021). The colony predation algorithm. Journal of Bionic Engineering, 18(3), 674-710. [Google Scholar] [CrossRef]
  37. Hussien, A. G., Amin, M., Wang, M., Liang, G., & Alsanad, A. (2020). Crow search algorithm: Theory, recent advances, and applications. IEEE Access, 8, 173548-173565. [Google Scholar] [CrossRef]
  38. Ahmadianfar, I., Heidari, A. A., Noshadian, S., Chen, H., & Gandomi, A. H. (2022). Info: An efficient optimization algorithm based on weighted mean of vectors. Expert Systems with Applications, 195, 116516. [Google Scholar] [CrossRef]
  39. Hashim, F. A., & Hussien, A. G. (2022). Snake optimizer: A novel meta-heuristic optimization algorithm. Knowledge-Based Systems, 242, 108320. [Google Scholar] [CrossRef]
  40. Yang, Y., Chen, H., Heidari, A. A., Gandomi, A. H. (2021). Hunger games search: Visions, conception, implementation, deep analysis, perspectives, and towards performance shifts. Expert Systems with Applications, 114864. 10.1016/j.eswa.2021.114864 [CrossRef]
  41. Hussien, A. G., Hassanien, A. E., Houssein, E. H., Amin, M., & Azar, A. T. (2019). New binary whale optimization algorithm for discrete optimization problems. Engineering Optimization, 52(6), 945-959. [Google Scholar] [CrossRef]
  42. Abdullahi, M., & Ngadi, M. A., Abdulhamid, S. I. M. (2016). Symbiotic organism search optimization based task scheduling in cloud computing environment. Future Generation Computer Systems, 56, 640-650. [Google Scholar] [CrossRef]
  43. Abualigah, L., Abd Elaziz, M., Hussien, A. G., Alsalibi, B., & Jalali, S. M. J. (2020). Lightning search algorithm: A comprehensive survey. Applied Intelligence, 51, 2353-2376. [Google Scholar] [PubMed] [CrossRef]
  44. Hussien, A. G., Amin, M., & Abd El Aziz, M. (2020). A comprehensive review of moth-flame optimisation: Variants, hybrids, and applications. Journal of Experimental & Theoretical Artificial Intelligence, 32(4), 705-725. [Google Scholar] [CrossRef]
  45. Li, S., Chen, H., Wang, M., Heidari, A. A., & Mirjalili, S. (2020). Slime mould algorithm: A new method for stochastic optimization. Future Generation Computer Systems, 111, 300-323. [Google Scholar] [CrossRef]
  46. Assiri, A. S., Hussien, A. G., & Amin, M. (2020). Ant lion optimization: Variants, hybrids, and applications. IEEE Access, 8, 77746-77764. [Google Scholar] [CrossRef]
  47. Heidari, A. A., Mirjalili, S., Faris, H., Aljarah, I., & Mafarja, M. (2019). Harris hawks optimization: Algorithm and applications. Future Generation Computer Systems, 97, 849-872. [Google Scholar] [CrossRef]
  48. Hussien, A. G., & Amin, M. (2022). A self-adaptive harris hawks optimization algorithm with opposition-based learning and chaotic local search strategy for global optimization and feature selection. International Journal of Machine Learning and Cybernetics, 13, 309-336. [Google Scholar] [CrossRef]
  49. Wang, S., Hussien, A. G., Jia, H., Abualigah, L., & Zheng, R. (2022). Enhanced remora optimization algorithm for solving constrained engineering optimization problems. Mathematics, 10(10), 1696. [Google Scholar] [CrossRef]
  50. Zheng, R., Hussien, A. G., Jia, H. M., Abualigah, L., & Wang, S. (2022). An improved wild horse optimizer for solving optimization problems. Mathematics, 10(8), 1311. [Google Scholar] [CrossRef]
  51. Elaziz, M. A., Hemedan, A. A., Ostaszweski, M., Schneider, R., & Lu, S. (2019). Optimization ace inhibition activity in hypertension based on random vector functional link and sine-cosine algorithm. Chemometrics and Intelligent Laboratory Systems, 190, 69-77. [Google Scholar] [CrossRef]
  52. Li, S., Fang, H., & Liu, X. (2018). Parameter optimization of support vector regression based on sine cosine algorithm. Expert Systems with Applications, 91, 63-77. [Google Scholar] [CrossRef]
  53. Nayak, D. R., Dash, R., Majhi, B., & Wang, S. (2018). Combining extreme learning machine with modified sine cosine algorithm for detection of pathological brain. Computers & Electrical Engineering, 68, 366-380. [Google Scholar] [CrossRef]
  54. Wang, J., Yang, W., Du, P., & Niu, T. (2018). A novel hybrid forecasting system of wind speed based on a newly developed multi-objective sine cosine algorithm. Energy Conversion and Management, 163, 134-150. [Google Scholar] [CrossRef]
  55. Dasgupta, K., Roy, P. K., & Mukherjee, V. (2020). Power flow based hydro-thermal-wind scheduling of hybrid power system using sine cosine algorithm. Electric Power Systems Research, 178, 106018. [Google Scholar] [CrossRef]
  56. Sahlol, A. T., Ewees, A. A., Hemdan, A. M., Hassanien, A. E. (2016). Training feedforward neural networks using sine-cosine algorithm to improve the prediction of liver enzymes on fish farmed on nano-selenite. 2016 12th International Computer Engineering Conference (ICENCO), pp. 35–40. Cairo, Egypt, IEEE.
  57. Sindhu, R., Ngadiran, R., Yacob, Y. M., Zahri, N. A. H., & Hariharan, M. (2017). Sine–cosine algorithm for feature selection with elitism strategy and new updating mechanism. Neural Computing and Applications, 28(10), 2947-2958. [Google Scholar] [CrossRef]
  58. Liu, G., Jia, W., Wang, M., Heidari, A. A., & Chen, H. (2020). Predicting cervical hyperextension injury: A covariance guided sine cosine support vector machine. IEEE Access, 8, 46895-46908. [Google Scholar] [CrossRef]
  59. Das, S., Bhattacharya, A., & Chakraborty, A. K. (2018). Solution of short-term hydrothermal scheduling using sine cosine algorithm. Soft Computing, 22(19), 6409-6427. [Google Scholar] [CrossRef]
  60. Oliva, D., Hinojosa, S., Elaziz, M. A., & Ortega-Sánchez, N. (2018). Context based image segmentation using antlion optimization and sine cosine algorithm. Multimedia Tools and Applications, 77(19), 25761-25797. [Google Scholar] [CrossRef]
  61. Tawhid, M. A., & Savsani, V. (2019). Multi-objective sine-cosine algorithm (MO-SCA) for multi-objective engineering design problems. Neural Computing and Applications, 31(2), 915-929. [Google Scholar] [CrossRef]
  62. Issa, M., Hassanien, A. E., Oliva, D., Helmi, A., & Ziedan, I. (2018). ASCA-PSO: Adaptive sine cosine optimization algorithm integrated with particle swarm for pairwise local sequence alignment. Expert Systems with Applications, 99, 56-70. [Google Scholar] [CrossRef]
  63. Khalilpourazari, S., & Pasandideh, S. H. R. (2020). Sine–cosine crow search algorithm: Theory and applications. Neural Computing and Applications, 32(12), 1-18. [Google Scholar]
  64. Chen, K., Zhou, F., Yin, L., Wang, S., & Wang, Y. (2018). A hybrid particle swarm optimizer with sine cosine acceleration coefficients. Information Sciences, 422, 218-241. [Google Scholar] [CrossRef]
  65. Nenavath, H., & Jatoth, R. K. (2018). Hybridizing sine cosine algorithm with differential evolution for global optimization and object tracking. Applied Soft Computing, 62, 1019-1043. [Google Scholar] [CrossRef]
  66. Elaziz, M. A., Oliva, D., & Xiong, S. (2017). An improved opposition-based sine cosine algorithm for global optimization. Expert Systems with Applications, 90, 484-500. [Google Scholar] [CrossRef]
  67. Chen, H., Wang, M., & Zhao, X. (2020). A multi-strategy enhanced sine cosine algorithm for global optimization and constrained practical engineering problems. Applied Mathematics and Computation, 369, 124872. [Google Scholar] [CrossRef]
  68. Chen, H., Heidari, A. A., Zhao, X., Zhang, L., & Chen, H. (2020). Advanced orthogonal learning-driven multi-swarm sine cosine optimization: Framework and case studies. Expert Systems with Applications, 144, 113113. [Google Scholar] [CrossRef]
  69. Guo, W., Wang, Y., Zhao, F., & Dai, F. (2019). Riesz fractional derivative elite-guided sine cosine algorithm. Applied Soft Computing, 81, 105481. [Google Scholar] [CrossRef]
  70. Gupta, S., & Deep, K. (2019). Improved sine cosine algorithm with crossover scheme for global optimization. Knowledge-Based Systems, 165, 374-406. [Google Scholar] [CrossRef]
  71. Gupta, S., & Deep, K. (2019). A hybrid self-adaptive sine cosine algorithm with opposition based learning. Expert Systems with Applications, 119, 210-230. [Google Scholar] [CrossRef]
  72. Long, W., Wu, T., Liang, X., & Xu, S. (2019). Solving high-dimensional global optimization problems using an improved sine cosine algorithm. Expert Systems with Applications, 123, 108-126. [Google Scholar] [CrossRef]
  73. Chen, H., Yang, C., Heidari, A. A., & Zhao, X. (2020). An efficient double adaptive random spare reinforced whale optimization algorithm. Expert Systems with Applications, 154, 113018. [Google Scholar] [CrossRef]
  74. Abd Elaziz, M. E., Ewees, A. A., Oliva, D., Duan, P., Xiong, S. (2017). A hybrid method of sine cosine algorithm and differential evolution for feature selection. Neural Information Processing, Cham: Springer International Publishing.
  75. Qu, C., Zeng, Z., Dai, J., Yi, Z., & He, W. (2018). A modified sine-cosine algorithm based on neighborhood search and greedy levy mutation. Computational Intelligence and Neuroscience, 2018, [Google Scholar] [PubMed] [CrossRef]
  76. Hussien, A. G., Heidari, A. A., Ye, X., Liang, G., Chen, H. et al. (2022). Boosting whale optimization with evolution strategy and Gaussian random walks: An image segmentation method. Engineering with Computers, 1–45. 10.1007/s00366-021-01542-0 [CrossRef]
  77. Su, H., Zhao, D., Yu, F., Heidari, A. A., & Zhang, Y. (2022). Horizontal and vertical search artificial bee colony for image segmentation of COVID-19 X-ray images. Computers in Biology and Medicine, 142, 105181. [Google Scholar] [PubMed] [CrossRef]
  78. Zeng, G. Q., Lu, K. D., & Dai, Y. X. Zhang, Z. J., Chen, M. R., . (2014). Binary-coded extremal optimization for the design of PID controllers. Neurocomputing, 138, 180-188. [Google Scholar] [CrossRef]
  79. Dong, R., Chen, H., Heidari, A. A., Turabieh, H., & Mafarja, M. (2021). Boosted kernel search: Framework, analysis and case studies on the economic emission dispatch problem. Knowledge-Based Systems, 233, 107529. [Google Scholar] [CrossRef]
  80. Wu, S. H., Zhan, Z. H., & Zhang, J. (2021). SAFE: Scale-adaptive fitness evaluation method for expensive optimization problems. IEEE Transactions on Evolutionary Computation, 25(3), 478-491. [Google Scholar] [CrossRef]
  81. Zhao, F., Di, S., Cao, J., Tang, J., & Jonrinaldi, . (2021). A novel cooperative multi-stage hyper-heuristic for combination optimization problems. Complex System Modeling and Simulation, 1(2), 91-108. [Google Scholar] [CrossRef]
  82. Cai, Z., Gu, J., Luo, J., Zhang, Q., & Chen, H. (2019). Evolving an optimal kernel extreme learning machine by using an enhanced grey wolf optimization strategy. Expert Systems with Applications, 138, 112814. [Google Scholar] [CrossRef]
  83. Yu, H., Cheng, X., Chen, C., Heidari, A. A., & Liu, J. (2022). Apple leaf disease recognition method with improved residual network. Multimedia Tools and Applications, 81(6), 7759-7782. [Google Scholar] [CrossRef]
  84. Lai, X., & Zhou, Y. (2020). Analysis of multiobjective evolutionary algorithms on the biobjective traveling salesman problem (1,2). Multimedia Tools and Applications, 79(1–3), 1-22. [Google Scholar] [CrossRef]
  85. Hu, J., Chen, H., Heidari, A. A., Wang, M., & Zhang, X. (2021). Orthogonal learning covariance matrix for defects of grey wolf optimizer: Insights, balance, diversity, and feature selection. Knowledge-Based Systems, 213, 106684. [Google Scholar] [CrossRef]
  86. Hu, J., Gui, W., Heidari, A. A., Cai, Z., & Liang, G. (2022). Dispersed foraging slime mould algorithm: Continuous and binary variants for global optimization and wrapper-based feature selection. Knowledge-Based Systems, 237, 107761. [Google Scholar] [CrossRef]
  87. Deng, W., Zhang, X., Zhou, Y., Liu, Y., & Zhou, X. (2022). An enhanced fast non-dominated solution sorting genetic algorithm for multi-objective problems. Information Sciences, 585, 441-453. [Google Scholar] [CrossRef]
  88. Tu, L., Yang, L. Z. (2013). Feature selection based on ant colony optimization for image classification. Applied Mechanics and Materials, 319, 337–342.
  89. Zhao, X., Li, D., Yang, B., Chen, H., & Yang, X. (2015). A two-stage feature selection method with its application. Computers & Electrical Engineering, 47, 114-125. [Google Scholar] [CrossRef]
  90. Deng, W., Liu, H., Xu, J., Zhao, H., & Song, Y. (2020). An improved quantum-inspired differential evolution algorithm for deep belief network. IEEE Transactions on Instrumentation and Measurement, 69(10), 7319-7327. [Google Scholar]
  91. Zhao, H., Liu, H., Xu, J., & Deng, W. (2019). Performance prediction using high-order differential mathematical morphology gradient spectrum entropy and extreme learning machine. IEEE Transactions on Instrumentation and Measurement, 69(7), 4165-4172. [Google Scholar]
  92. Shi, B., Ye, H., Zheng, L., Lyu, J., & Chen, C. (2021). Evolutionary warning system forCOVID-19 severity: Colony predation algorithm enhanced extreme learning machine. Computers in Biology and Medicine, 136, 104698. [Google Scholar] [PubMed] [CrossRef]
  93. Sun, Y., Xue, B., Zhang, M., & Yen, G. G. (2020). Evolving deep convolutional neural networks for image classification. IEEE Transactions on Evolutionary Computation, 24(2), 394-407. [Google Scholar] [CrossRef]
  94. Qiao, K., Liang, J., Yu, K., Yuan, M., & Qu, B. (2022). Self-adaptive resources allocation-based differential evolution for constrained evolutionary optimization. Knowledge-Based Systems, 235, 107653. [Google Scholar] [CrossRef]
  95. Liang, J., Ban, X., Yu, K., Qu, B., & Qiao, K. (2021). Differential evolution with rankings-based fitness function for constrained optimization problems. Applied Soft Computing, 113, 108016. [Google Scholar] [CrossRef]
  96. Gao, D., Wang, G. G., & Pedrycz, W. (2020). Solving fuzzy job-shop scheduling problem using de algorithm improved by a selection mechanism. IEEE Transactions on Fuzzy Systems, 28(12), 3265-3275. [Google Scholar] [CrossRef]
  97. Wang, G. G., Gao, D., & Pedrycz, W. (2022). Solving multiobjective fuzzy job-shop scheduling problem by a hybrid adaptive differential evolution algorithm. IEEE Transactions on Industrial Informatics, 18(12), 8519-8528. [Google Scholar] [CrossRef]
  98. Hu, K., Ye, J., Fan, E., Shen, S., & Huang, L. (2017). A novel object tracking algorithm by fusing color and depth information based on single valued neutrosophic cross-entropy. Journal of Intelligent & Fuzzy Systems, 32(3), 1775-1786. [Google Scholar] [CrossRef]
  99. Hu, K., He, W., Ye, J., Zhao, L., & Peng, H. (2019). Online visual tracking of weighted multiple instance learning via neutrosophic similarity-based objectness estimation. Symmetry, 11(6), 832. [Google Scholar] [CrossRef]
  100. Mirjalili, S. (2016). SCA: A sine cosine algorithm for solving optimization problems. Knowledge-based Systems, 96, 120-133. [Google Scholar] [CrossRef]
  101. Vaishnav, P. K., Sharma, S., & Sharma, P. (2021). Analytical review analysis for screening COVID-19 disease. International Journal of Modern Research, 1(1), 22-29. [Google Scholar]
  102. Li, C., Li, J., Chen, H., & Heidari, A. A. (2021). Memetic harris hawks optimization: Developments and perspectives on project scheduling and qos-aware web service composition. Expert Systems with Applications, 171, 114529. [Google Scholar] [CrossRef]
  103. Ren, H., Li, J., Chen, H., & Li, C. (2021). Stability of salp swarm algorithm with random replacement and double adaptive weighting. Applied Mathematical Modelling, 95, 503-523. [Google Scholar] [CrossRef]
  104. Wilcoxon, F. (1992). Individual comparisons by ranking methods. In: Kotz, S., Johnson, N. L., Breakthroughs in statistics: Methodology and distribution, pp. 196–202. New York, NY: Springer.
  105. Çelik, Y., & Kutucu, H. (2018). Solving the tension/compression spring design problem by an improved firefly algorithm. IDDM, 1(2255), 1-7. [Google Scholar]
  106. Deb, K. (1991). Optimal design of a welded beam via genetic algorithms. AIAA Journal, 29(11), 2013-2015. [Google Scholar] [CrossRef]
  107. Moss, D. R. (2004). Pressure vessel design mannual. Elsevier. 10.2514/3.10834 [CrossRef]
  108. Morton, S., & Webber, J. (1994). Optimal design of a composite i-beam. Composite Structures, 28(2), 149-168. [Google Scholar] [CrossRef]
  109. Gupta, V. K., Surendra Kumar, S., Anupriya., & Ramesh Singh, R. (2022). Crime tracking system and people’s safety in India using machine learning approaches. International Journal of Modern Research, 2(1), 1-7. [Google Scholar]
  110. Shukla, S. K., Gupta, V. K., Joshi, K., Gupta, A., & Singh, M. K. (2022). Self-aware execution environment model (SAE2) for the performance improvement of multicore systems. International Journal of Modern Research, 2(1), 17-27. [Google Scholar]

Appendix A. Tension/compression spring design problem

Minimize: f(x)=(x3+2)x2x12

Subject to: g1(x)=1(x23x3/71,785x14)0

g2(x)=(4x22x1x2/12,566(x2x13x14)+(1/5108x12))100

g3(x)=1(140.45x1/x22x3)0

g4(x)=(x2+x1)/1.510,

Variable Range

0.05x12.00

0.25x21.30

2.00x315.00

Appendix B. Welded beam design problem

Minimize: f1(x)=1.10471x(1)2x(2)+0.04811x(3)x(4)(14.0+x(2))

Subject to: g1(x)=τ13600

g2(x)=σ30000

g3(x)=x(1)x(4)

g4(x)=6000p

Variable Range

0.125x15

0.1x210

0.1x310

0.125x45

Appendix C. Pressure vessel design problem

Minimize: f(x)=0.6224x1x3x4+1.7781x2x32+3.1661x12x4+19.84x12x3

Subject to: g1(x)=x1+0.0193x

g2(x)=x2+0/00954x30

g3(x)=πx32x4(4/3)πx33+1,296,0000

g4(x)=x42400

Variable Range

0xi100,i=1,2

0xi200,i=3,4

Appendix D. I-beam design problem

Minimize: f(x)=5000tw(h2tf)312+btf36+2btf(htf2)2

Subject to:

g=2btw+tw(h2tf)0

Variable Range

10x150

10x280

0.9x35

0.9x45

Figure S1: Balance analysis of RDSCA and SCA

Table S1: CEC 2017 benchmark functions

Table S2: The calculated p-values from the Wilcoxon signed-rank test

Table S3: Results of TCSD problem

Table S4: Results of WBD problem

Table S5: Results of PVD problem

Table S6: Results of IBD problem


Cite This Article

Hussien, A. G., Liang, G., Chen, H., Lin, H. (2023). A Double Adaptive Random Spare Reinforced Sine Cosine Algorithm. CMES-Computer Modeling in Engineering & Sciences, 136(3), 2267–2289. https://doi.org/10.32604/cmes.2023.024247


cc This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 2078

    View

  • 822

    Download

  • 0

    Like

Share Link