1. Introduction
An iterative series of task sequences, data selection and pretreatment, mining algorithm selection, data mining, pattern evaluation, and knowledge presentation make up knowledge discovery in databases (KDD) [1,2,3]. The main objective of data preprocessing, as the initial stage in KDD, is to prepare datasets for use by data mining algorithms [4]. However, as information science progresses, the dimensionality of datasets increases dramatically, affecting the performance of clustering and classification approaches [5,6,7]. High-dimensional datasets also have data redundancy, performance deterioration, and a more extended period to build models [8,9,10]. These limitations have given rise to more difficulties in data analysis. Feature selection (FS) is frequently used as a preprocessing approach in the data mining process to determine the best subset of features from all available feature sets [11,12,13]. It eliminates irrelevant and redundant features, simplifies clustering and classification, enhances accuracy, and the problem of symmetry and asymmetry is also solved to a certain extent [14,15,16].
While some feature selection methods can solve the problem exactly for linear models only with promising results, exact methods can only handle hundreds or thousands of features at best. Another shortcoming of most feature selection methods is that they arbitrarily seek to identify only one solution to the problem. However, in practice, there are often multiple predictive or even information-equivalent solutions. Especially in fields where there is inherent redundancy in the underlying problem, as in molecular biology, there are often multiple solutions [17,18].
There are many ways to solve the FS problem, which can generally be divided into the following three categories: filters, wrappers, and embedded methods [19,20]. For the filtering method, the features in a given feature set are first sorted according to a series of criteria. Then the features with a higher ranking are formed into feature subsets [21]. Although the obtained feature subset is not necessarily the optimal subset, its calculation efficiency is very high, so this method is often used for high-dimensional FS problems. Representative filtering methods include minimum redundant F-score criteria [22], maximum correlation (mRMR) [23], Gini index [24], and correlation coefficient [25]. Wrapper approaches often use a predetermined learning process that is evaluated using a subset of features [26]. In most circumstances, wrapper methods outperform filter approaches that are not dependent on any learning mechanism. The embedded methods are embedded in the learning algorithm, and a subset of features can be obtained when the training process of the classification algorithm is completed [27]. The embedded method can solve the problem of excessive redundancy in the results of the filter methods based on feature sorting. It can also solve the problem of excessive time complexity of the wrapper methods, which is a compromise between the filter and wrapper [28,29,30].
Various methods for discovering optimal feature subsets have arisen in the wake of the wrapper-based method, including heuristic search, complete search, greedy search, and random search, to mention a few [31,32,33,34]. However, most of these methods suffer from local optima and expensive computational costs due to the use of greedy search methods [35,36]. Over the past three decades, Evolutionary Algorithms (EA) have been very reliable in solving various optimization problems, such as image processing [37], intrusion detection [38], path planning [39], particle filtering [40], production scheduling [22], support vector machines [41], home healthcare [42], wireless sensors [43], and neural network models [44].
Due to its capabilities in seeking competitive solutions employing tactics that perform well in exploration, EA has recently gained much attention in tackling FS challenges [45,46,47]. These approaches include Genetic Algorithm (GA) [48], Particle Swarm Optimization (PSO) [49], and White Shark Optimizer (WSO) [50]. A comprehensive review of nature-inspired FS techniques can be found in [22], and a detailed analysis of EA to FS can be found in [51]. Here are some examples.
Based on GA, the K-Nearest-Neighbors (K-NN) approach for diagnosing patient diseases was proposed in [52]. It used a hybrid genetic algorithm to perform efficient feature selection. The K-NN algorithm is utilized to diagnose lung cancer after an experimental technique was employed to determine the ideal value of K. In [53], by minimizing the numerous objectives of the FS, a non-dominated sorting genetic algorithm (NSGA) is employed to solve the multi-objective optimization problem (MOP). Recently, Xue et al. [54] proposed a multi-objective binary genetic algorithm called MOBGA-AOS with five crossover operators.
Based on PSO, Song et al. [55] proposes a K-NN and mutual information-based bare-bones PSO (BBPSO) feature selection algorithm. The adaptive flip mutation operator intends to assist particles in breaking out from the optimal local solution. In [56], a high-dimensional FS problem is solved using a multi-stage hybrid FS algorithm (HFS-C-P) based on PSO. Recently, Li et al. [57] proposed an improved Sticky Binary PSO (ISBPSO) algorithm for FS.
Grey Wolf Optimizer (GWO) and Sparrow Search Algorithm (SSA) are other EAs that also investigate solving the FS problem. Jangir et al. [58] proposed that non-dominated sorting GWO (NSGWO) is used to perform FS to improve the categorization of cervical lesions by reducing the number of textural features while increasing the classification accuracy. Recently, Sathiyabhama et al. [59] proposed an FS framework based on GWO and a rough set method called GWORS for finding salient features from extracted mammograms. Based on SSA, Chen et al. [60] proposed a spark-based improved SSA (SPISSA) used to search feature subsets on intrusion detection datasets.
In addition, applying FS technology based on EA in detecting COVID-19 patients is also becoming more extensive. da Silva et al. [61] combined single models such as Stereo Regression, Quantile Random Forest, K-NN, Bayesian Regression Neural Network, and Support Vector Regression with Variational Mode Decomposition (VMD) to create a hybrid model to forecast COVID-19 cases in Brazil. The VMD-based model proved to be one of the most effective strategies for forecasting COVID-19 cases five days in advance. Dey et al. [62] presented a hybrid model. To begin, they extracted several characteristics from the COVID-19-affected lungs. The Manta-Ray Foraging-based Golden Ratio Optimizer (MRFGRO), a hybrid meta-heuristic FS technique, is then presented to pick the most critical subset of characteristics. Although the findings show that the proposed strategy is quite effective, the model was only tested on the CT scan dataset. Shaban et al. [63] proposes Distance-Biased Naive Bayes (DBNB), a new approach for detecting COVID-19-infected patients. DBNB picks the most informative characteristics for identifying COVID-19 patients through a novel FS technique called Advanced PSO (APSO). APSO is a hybrid strategy that uses filter and wrapper approaches to offer accurate but crucial classification features.
In conclusion, many FS approaches employ EA to avoid increasing computing complexity in the high-dimensional dataset. These algorithms use primitive mechanisms and operations to solve an optimization problem and iteratively search for the optimal solution. Nonetheless, the No Free Lunch (NFL) [64] theorem states that existing procedures can constantly be improved. Moreover, there is currently insufficient research in the literature to solve the FS problem using the SSA, motivating us to suggest a variant of SSA for FS in Section 3.
The SSA is a new and well-organized EA that can be used in different areas for solving optimization problems, such as brain tumor detection [65], parameter identification [66], configuration network [67], and fault diagnosis of wind turbines [68]. However, SSA still has the problem of being easy to fall into the local optimum, and so far, the application of the SSA for FS is very scarce [69]. Motivated by the above analysis, we propose a Tent Lévy Flying Sparrow Search Algorithm (TFSSA) in this paper to increase the capability of SSA in confronting FS challenges, where the main contributions are summarized below.
A TFSSA is proposed for feature selection problems, and it is utilized to solve a COVID-19 case study.
An improved Tent chaos strategy, Lévy flights (LFs) mechanism, and Self-adaptive hyper-parameters are integrated into TFSSA to improve SSA’s exploratory behavior and perform well in the CEC2020 benchmark function.
A comprehensive comparison of TFSSA and nine different algorithms for feature selection problems is undertaken in nine aspects.
The proposed TFSSA’s improved searching capabilities are tested on 21 well-known feature selection datasets with excellent results.
The remainder of this paper is organized as follows: Section 2 presents the background of the SSA. The proposed TFSSA is described in Section 3. Section 4 presents the proposed TFSSA algorithms for FS, while the experimental results with discussions are reported in Section 5. Section 6 demonstrates the adoption of the proposed TFSSA in a COVID-19 application. Finally, we conclude this paper in Section 7.
2. SSA
2.1. Background of SSA
The literary representation of various animal, insect, and bacterial populations in nature provides a fascinating field of study for diverse scientific researchers. By simulating the foraging and reproduction behaviors of animal, insect, or bacterial communities, researchers draw inspiration from it to various abstract swarm intelligence and evolutionary behaviors into quantifiable vital indicators, which, in turn, form mathematical models that can be used to address various realities questions. The introduction of many meta-heuristic algorithms has greatly enriched optimization techniques and provided new tools for exploring the concepts and mechanisms of the biological world from another perspective. Based on the above, In 2020, Xue Jiankai proposed the SSA to enhance optimization technology and decode the complexity involved in the process [70].
2.2. Advantages of SSA from Other EA
SSA differs from other EA with several advantages. First, SSA does not update rules according to simulated social creatures’ step size but sets up rules according to its algorithm mechanism. It can handle various optimization issues with only four proprietary parameters to change. Second, SSA’s mathematical model makes it suited for resolving a range of engineering optimization issues, particularly those that involve high dimensions. Third, SSA’s resilience and simplicity allow it to identify global solutions to complicated optimization problems with high convergence rates. Fourth, SSA has gradually become a strong competitor with a broad interest in developing low-cost and robust solutions to actual optimization issues.
2.3. Rule Design
The classical SSA is primarily motivated by a sparrow population’s foraging behavior. It is a search algorithm with high optimization and efficiency capabilities [70,71,72,73]. For simplicity, the biology of sparrow populations during foraging is idealized and normalized as the following behaviors.
(1). Producers (leaders) have access to plentiful food sources and are responsible for ensuring that all scroungers (followers) have access to foraging sites.
(2). Some sparrows will be chosen as patrollers (guards). When patrollers come across a predator, they will sound an alarm. When the safety threshold is exceeded, the producer must direct the scroungers (followers) to other safe regions.
(3). Sparrows that can discover a better food source earn more energy and are promoted to producers (leaders). At the same time, hungry scroungers (followers) are more likely to fly elsewhere to forage to gain more energy, and the producer-to-forager ratio remains steady.
(4). Scroungers (followers) hunt for food after the finest producers (leaders). Simultaneously, certain predators may observe producers (leaders) and steal food.
(5). When threatened, sparrows near the flock’s edge moved swiftly to a safe region, while sparrows in the center of the flock moved randomly to approach other sparrows in the safe area.
2.4. Algorithm Design
The algorithm design of SSA is summarized in the following main steps.
Step 1: Parameter initialization, which includes the number of sparrows (N), the number of producers (), the number of patrollers (), the number of guards (), is a subset of N, the safety threshold (), the warning value (), and the maximum iterations (). The following matrix can be used to depict the initial position of sparrows:
(1)
where X is the initial location of the sparrow population, N is the number of sparrow populations, D is the variable dimension of the problem to be optimized, , and . The fitness value () of the sparrow population is represented by vectors as follows:(2)
where the value of f represents the fitness value of the sparrow individual, and .Step 2: According to the design of (1) and (2) in Section 2.3, the producer usually has a better fitness value, and it has a higher priority to capture food than other individuals in the search process. Producers are responsible for finding food for the entire population and providing foraging directions for others. As a result, producers have access to a broader search space than scavengers. The producer position of the sparrow population is updated in each iteration as follows:
(3)
where ; T_max is the maximum number of iterations; t indicates the current iteration; represents the value of the ith sparrow at iteration t; represents the safety threshold, and represents the warning value; is a random number; L shows a vector of ; Q is a random number; . When , it means that there are no predators around, and it is safe. The producer turns on the wide-area search mode, and the entire population is in a safe foraging state. , which means the patrollers find the danger and raise the alarm to warn the companions that there are predators around, and all the sparrows fly to other safe areas to avoid the danger.Step 3: According to rules (3) and (4) in Section 2.3, some followers keep a closer eye on leaders (producers). When the followers spot the producers who have located the food, they will promptly leave their current place to collect the food. The scrounger’s position update formula is as follows:
(4)
where , A represents a vector of , where each element is randomly assigned . represents the current global worst position. is the current optimal position of the producer. When , the ith scavenger with a lower fitness value did not get food and was in a state of starvation and had to fly to other places to find food at this time.Step 4: We hypothesized that the patrollers made up 10–20% of the population in our simulations, where is the number of guards (patrollers). These sparrows’ initial placements are generated at random. The mathematical model is expressed as follows according to rule (5) in Section 2.3:
(5)
where represents the current global optimal position; is a random number that obeys the standard Gaussian distribution; is the smallest constant to avoid division by zero errors; is a random number; the current fitness value of the sparrow is ; and and represent the current global best fitness value and worst fitness value, respectively.For simplicity, represents the sparrow at the edge of the group. represents the center of the population location around which it is safe. indicates that sparrows in the middle of the population need to approach other sparrows because they know the danger. K is the step size control coefficient, which indicates the direction in which the sparrow moves.
Step 5: Calculate, compare, and update the current position of the sparrow population, and sort and update the fitness values.
Step 6: Repeat Steps 2 through 5 until the maximum number of iterations () has been reached, at which point the best position () and best solution () will be output.
Algorithm 1 demonstrates the algorithmic structure of the classic SSA.
Algorithm 1 SSA |
Input: |
The number of sparrows(N) |
The number of producers() |
The number of guards() |
The warning value() |
The maximum iterations() |
Output: |
The best position() |
The best solution() |
1: t ← 0; |
2: while (t < T_max) do |
3: Calculate and update the , , and ; |
4: for each leaders do |
5: The location of leaders(producers) is updated using Equation (3); |
6: end for |
7: for each followers do |
8: The location of followers(scroungers) is updated using Equation (4); |
9: end for |
10: for each patrollers do |
11: The location of patrollers is updated using Equation (5); |
12: end for |
13: Find the current new location ; // If the new location is better than before, update it. |
14: Rank the ; |
15: t ← t + 1; |
16: end while |
17: return , . |
3. The Proposed Algorithm
This section delineates the TFSSA. As mentioned in Section 2, although the SSA has the advantages of faster convergence and more vital optimization-seeking abilities, the original SSA, like other traditional EAs, is more subject to the population’s poor diversity and falls into a local optimum. The placements of the sparrows in the solution space are randomly distributed, and a random walk method [66] is used when no nearby sparrows surround the current individual. This mode delays the convergence trend and reduces the convergence accuracy for a limited number of iterations. Our proposed algorithm aspires to improve SSA’s complete optimization performance and address these shortcomings.
3.1. Initialized Population
Initialization is a severe step in the meta-heuristic algorithm and affects convergence speed and solution accuracy. The primary motivation of the most advanced initialization methods is to cover the search space as evenly as possible based on generating a small initial population. However, these methods are affected by the dimension disaster, high computational cost, and sensitivity to parameters, which ultimately reduce the convergence speed of the algorithm [73,74].
The efficiency of EA is greatly affected by chaotic mapping, which has the advantages of uniform ergodicity, sensitivity to initial values, and fast search speed. Using the randomicity, ergodicity, and regularity of chaotic variables to solve optimization problems can make the algorithm jump out of local optimization, maintain population diversity, and improve the global search ability to a certain extent. However, different chaotic maps significantly impact the chaotic optimization process. Various scholars have introduced chaos mapping and chaos search into EA, trying to improve the problem of falling into local optimums in the latter period and improve the convergence speed and accuracy of the algorithm. The chaotic map used more in the literature is the Logistic map. Still, its value probability is high in the two ranges [0, 0.1] and [0.9, 1], and the optimization speed is affected by the uneven Logistic traversal, so the algorithm’s efficiency will be significantly reduced. Many papers have pointed out that the Tent map has better ergodic uniformity and faster convergence speed than the logistic map and have further proved that the Tent map can be used as a chaotic sequence to generate optimization algorithms through strict mathematical reasoning. The Tent mapping expression is shown in Equation (6).
(6)
Equation (6) after Bernoulli shift transformation is as follows:
(7)
where is the modulo function. Tent mapping has the advantages of randomness, consistency, and orderliness, and it is a standard method for scholars to find the optimal solution [72,75]. On the other hand, a chaotic Tent map has flaws, such as a short period and unstable period points [76]. Therefore, the Tent chaos map is enhanced by the , as given in Equation (8), to avoid slipping into a tiny period or an unstable periodic point [74].(8)
where in the current experiments, × , and N represents the population of sparrows. Equation (8) after Bernoulli shift transformation is as follows:(9)
where is the modulo function, × , and N represents the population of sparrows.Therefore, in TFSSA, Equations (1) and (9) in the traditional SSA are replaced by Equation (8) to increase the sparrow population diversity. At this time, the improved Tent chaotic sequence is introduced based on the original SSA to initialize the sparrow population. Although the algorithm not only retains the randomness of the initial individuals but also improves the population diversity at the initial stage, it still cannot guarantee that the population diversity will still have the same degree later. However, in subsequent experiments, it was found that the population’s diversity is not well guaranteed, and the scavengers constantly hop around the producers, making the algorithm fall into local optimization to a large extent in the algorithm’s later stages. In this case, we consider introducing the LF mechanism to improve the algorithm’s performance further.
3.2. LF Mechanism
“LFs” are named after the French mathematician Paul·Lévy (1886–1971), who first proposed the concept in 1937. LFs try to strengthen the optimization process with diversity and universality, which helps the algorithm find the search location effectively and avoid local minima. Therefore, LFs embed in the SSA mechanism to improve the overall optimization efficiency. The foraging activities of most animals are also characteristic of LFs, for example, the routes of plankton, termites, bumblebees, birds, and primates. LFs seem to be a common law for creatures to survive in a resource-scarce environment when foraging has similar patterns. The trajectory of human beings when traveling and shopping also belongs to LFs.
It can be seen from the SSA rule design (3) that when the producer’s food does not have enough temptation, hungry scavengers may fly to other places to look for food. However, according to the SSA rule design (4), scavengers mainly search for food from the producer and go elsewhere to look for food. Generally, they only search for food within a relatively close range of the producer. Therefore, most sparrows may only move around areas with poor solution quality. On the other hand, for each iteration, the individual sparrow will move indiscriminately to the sparrow (producer) whose food is better than his own. This situation increases the algorithm’s complexity and leads to low convergence accuracy and a higher possibility of falling into the optimal local solution. Random numbers obeying the Lévy distribution have the characteristics of short-distance walking and long-distance jumping, which will significantly improve the disadvantage of hungry sparrows (scavengers) that only search for food within a relatively close range of producers.
In summary, this part combines the LF strategy and the inertia weight factor into the classic SSA to improve its ability to expand the search scope and avoid local optimization. In this way, TFSSA can locate the optimal global solution more effectively. Equations (10)–(13) describe this mechanism. Equation (10) can be used to express the Lévy distribution [77].
(10)
where is a stability index, and , the is a Gaussian distribution. The inertia-weighting factor is expressed by Equation (11).(11)
then, the sparrow’s position, , is mutated by the random roulette strategy in Equation (12). If rand > ,(12)
else the is also changed by Equation (13).(13)
where L(α) is a number chosen randomly from the Lévy distribution. This part mainly combines the LF strategy with classic SSA and uses LF characteristics to improve its ability to expand the search scope and avoid local optimization. LFs can optimize the diversity of search agents, enabling the algorithm to explore search locations and avoid local minima effectively. The combination of LFs and the SSA algorithm improves the population diversity to a certain extent and enhances the robustness and global optimization capability of the SSA algorithm. However, in many experiments, it is found that the occasional long-distance jump of LFs has no significant impact on the final performance of the algorithm, as expected. Because of its poor performance in the CEC2020 benchmark function, we are considering further improving the algorithm from the manufacturer’s location formula. We use super adaptive parameters in the next section to update the producer location and improve the global search capability.3.3. Self-Adaptive Hyper-Parameters
In the rule design of SSA in Section 2.3, SSA mainly divides the sparrow population into producers (leaders) and scavengers (followers). Producers need more search space to find food sources, while scavengers mainly follow producers to find food. Therefore, the global search capability of the original SSA is highly dependent on the search scope of the producer.
In Equation (3), means that there are no predators at present, and the producer (leader) opens the wide-area search mode. In this mode, the location update of producers (leaders) is mainly affected by . When in Equation (3) gets a large random value, the value of will gradually decrease from (0, 1) to (0, 0.4) as i becomes larger. To sum up, we expand the search range of producers by adaptive control factors. The adaptive control factors are shown below in Equation (14).
(14)
where t is the current iteration number; is the initial weight; c is the adaptive factor of w, which can be modified depending on the actual problem; and w is the initial weight. According to the subsequent sensitivity analysis, the performance of the TFSSA is relatively stable and achieves its best classification accuracy in most cases of datasets when c is 0.9. Therefore, in our current research, c is set to 0.9 to maintain w at a low value, enhancing the global search capability and broadening the producers’ search scope. The original producer position is updated from Equation (3) to Equation (15).(15)
In addition, to detect and warn companions to avoid predators in time during the foraging process, one-tenth to one-fifth of sparrows are selected as guards, also called patrollers. When the patroller realizes the danger and issues an alarm, the entire population will immediately conduct anti-predation behaviors, thereby improving the entire population’s predation ability and risk-prevention capabilities. In other words, the presence of patrollers can help sparrow populations achieve better SSA solutions. When the number of patrollers is large, it is beneficial to improve the global optimization ability of sparrows. However, as the number of patrollers is reduced, it aids in accelerating SSA convergence. Therefore, this paper proposes an adaptive update formula for the number of patrollers, aiming to improve the algorithm’s performance by adjusting the number of patrollers, as shown by Equation (16), the formula can non-linearly reduce during the iteration.
(16)
where represents the number of patrollers; represents the maximum number of patrollers; represents the minimum number of patrollers; the function is used to round values; t is the current iteration; and is the maximum number of iterations. The in the original SSA is set to be chosen at random from 10–20% of the sparrow population. Equation (16) in this paper replaces the random selection behavior of the number of patrollers in the original SSA and better balances the algorithm convergence speed and global optimization ability. When all the sparrows find the optimal solution, this paper mutates the optimal sparrow individuals again to improve the global convergence accuracy further.3.4. Optimal Individual Mutation by -Tent Chaos
The original SSA is prone to fall into local extrema in the later iterations. To solve this problem, the optimal individual position is perturbed in each iteration, and only one individual is randomly mutated in each iteration. That is, when the sparrow finds the optimal solution, the enhanced Tent chaos is used to mutate the optimal sparrow individual, which further improves the global convergence accuracy and optimizes the shortcomings of the original algorithm in global search and local search [78]. Therefore, in TFSSA, the optimal sparrow individuals are changed by Equations (17) and (18).
(17)
if < r, then the optimal sparrow position is updated by Equation (18).(18)
where ψ-Tent can be calculated by Equation (8). The overall flow of TFSSA is in Algorithm 2.Algorithm 2 TFSSA |
Input: |
The number of sparrows(N) |
The number of producers() |
The number of guards() |
The safety threshold() |
The warning value() |
The maximum iterations() |
Output: |
The best position so far() |
The best solution so far() |
1: Initialize a flock of sparrows’ location X // Pretreatment by Equations (8) and (9). |
2: t ← 0; |
3: while (t < T_max) do |
4: Rank the fitness vaule using Equation (2); |
Find the and ; |
Update the ← a random value in [0, 1], and calculate the using Equation (17). |
5: for each leaders do |
6: The location of leaders(producers) is updated using Equation (15); // The original producer position is updated from Equation (3) to Equation (15). |
7: end for |
8: for each followers do |
9: The location of followers(scroungers) is updated using Equation (4); |
10: end for |
11: for each patrollers do |
12: The location of patrollers is updated using Equation (5); // The is updated using Equation (16). |
13: end for |
14: Update and . |
15: for do |
16: if ( > ) then |
17: The is updated using Equation (12). // indicates the inertia weighting factor. |
18: else |
19: The is mutated using Equation (13). |
20: end if |
21: end for |
22: Update and . |
23: Calculate the r using Equation (17). |
24: if ( < r) then |
25: ← ; // The is mutated using Equation (18). |
26: end if |
27: Rearrange all of the population’s in ascending order. |
28: ← ; // Update . |
29: ← ; // Update . |
30: t ← t + 1; |
31: end while |
32: return , . |
3.5. Computational Complexity Analysis
This subsection uses the well-known Big-O notation to present the proposed TFSSA’s time and computational space complexity. Although the proposed TFSSA and SSA have the same time complexity of O(N) and space complexity of O(D × N), TFSSA performs better than SSA in the sequence experiments.
3.5.1. Time Complexity Analysis
The time complexity depends on the size of the sparrow population (N), the dimension of the problem (D), the maximum number of iterations (T_max), the number of producers (PN) and scroungers (N-PN) along with the number of patrollers (GN). The time complexity of stage (1) in SSA is O(D × N), the time complexity of stages (2) and (3) is O(D × N), and the time complexity of stage (4) is O((PN + N-PN + GN) × T_max × D), which is O(N × T_max × D); hence the total time complexity of SSA is O(N). The proposed TFSSA mainly includes the stages shown in Figure 1. The computational cost of TFSSA is primarily different from that of stage (4) in comparison to SSA. TFSSA has a computational complexity of O(T_max × D × 3N) when it comes to the sparrow’s location updating phase. To summarize, the proposed TFSSA and classical SSA have a time complexity of O(N).
3.5.2. Space Complexity Analysis
The space complexity of TFSSA relative to the amount of memory space depends on the number of sparrows and the dimensions of the problem. This determines the total amount of memory space required for the input values that the proposed TFSSA uses for execution. Therefore, without considering the auxiliary space, the space complexity of TFSSA and SSA is O(D × N).
4. TFSSA Applied for FS
In this section, we introduce the application of TFSSA in classification tasks. In the novel algorithm, We start by discretizing the initial position of each sparrow in the group generated by the chaotic initialization of -Tent to each dimension. Then, we set the fitness function utilized in TFSSA to evaluate individual sparrow placements. Finally, the process iterates until the stop criterion is met and the optimal feature subset’s feature space is obtained. In the following, we detail the application in FS of the proposed methods.
4.1. Initialization
The initialization stage is the first step of the EA, in which, according to Equations (8) and (9), a sparrow individual of N is generated through the chaotic initialization of -Tent. In this study, we try to identify the significant 1 value and reject the other feature 0 value. Before starting the fitness evaluation process, according to Equations (8) and (9) and Figure 2, we first discretize the initial position of each sparrow in the group to the position on each dimension; that is, 0 (not selected) or 1 (selected), and convert it to a random binary value (between 0 and 1).
4.2. Fitness Evaluation
In this part, the TFSSA is exploited in FS for classification problems. The different feature combinations for a feature vector of size would be , which is a massive space of features to be searched thoroughly. As a result, TFSSA is utilized to choose the optimal feature subset’s feature space. Equation (19) shows the fitness function utilized in TFSSA to evaluate individual sparrow placements.
(19)
where is the error rate for the classifier of condition attribute set, denotes the ratio of chosen features to total features, and = 1 − .K-Nearest Neighbors (K-NN) [79] is a popular classification method that may be used to evaluate fitness functions as a simple candidate classifier. The smallest distance between the query instance and the training examples determines the K-NN classifier. A crucial characteristic of wrapper techniques in FS is the use of the classifier as a guide to the FS activity. The following three primary items can be used to classify wrapper-based feature selection: (1) Method of classification. (2) Criteria for evaluating features. (3) Search method. As demonstrated in Equation (19), TFSSA is employed as a search strategy that may adaptively explore the feature space to maximize the feature evaluation criterion. A sparrow’s location in the search space reflects one feature combination or solution since each dimension represents a different feature combination or solution.
4.3. Termination
In each iteration, the position of sparrows (producers, scavengers, patrollers) is updated (refer to Algorithm 2), and the continuous value of the position vector is recorded after each iteration for future use in the continuous position update of the entire constant iteration. Next, the process iterates until the stop criterion is met, that is, the maximum number of function evaluations in this study.
5. Experimental
In this section, we introduce the evaluation of TFSSA in benchmark functions and multi-perspective analysis. Then, we discuss the performance of the proposed algorithm in FS.
5.1. Evaluation of TFSSA
The CEC2020 benchmark suite is selected to evaluate the effectiveness and superiority of the proposed algorithm, TFSSA, and compare it with seven other algorithms, including the Artificial Bee Colony Algorithm (ABC), PSO, Competitive Swarm Optimizer (CSO), DE, SSA, Optimal Foraging Algorithm (OFA), and Success History-based Adaptive Differential Evolution (SHADE). The reasons for choosing the CEC2020 benchmark suite are discussed in Section 5.1.1.
5.1.1. Benchmark Functions
CEC benchmarks are the most widely used benchmark problems and have been used by many research scientists to test their algorithms. The most popular single-objective optimization test function set includes CEC2005 [80], CEC2008 [81], CEC2010 [82], CEC2013 [83], CEC2014 [84], CEC2017 [85], CEC2020 [86], and the Single-Objective optimization problem (SOP) [87]. The single-objective optimization algorithm is the basis for building more complex methods, such as multi-objective, super multi-objective, multi-modal multi-objective, niche, and constrained optimization algorithms. Therefore, it is crucial to improve single-objective optimization algorithms because they will also impact other areas. These algorithm improvements, to some extent, depending on the feedback of experiments conducted using single objective benchmark functions, which are basic components of more complex tasks. With the improvement in algorithms, researchers must develop more challenging functions to adapt to new problems. The interaction between methods and problems promoted progress, and CEC2020, developed by researchers further, promoted this symbiotic relationship.
The improved methods and problems sometimes need to update the traditional test standards, and the traditional test standards (such as SOP) can not guarantee enough persuasiveness when facing new, improved algorithms. Therefore, this paper uses the classical test function set CEC2020 to test the comprehensive performance of the proposed algorithm TFSSA. CEC2020 includes one unimodal function (CEC2020_F1), three basic functions (CEC2020_F2–CEC2020_F4), three mixed functions (CEC2020_F5–CEC2020_F7), and three synthesis functions (CEC2020_F8–CEC2020_F10), as shown in Table 1. The MATLAB and C code for the CEC2020 test suite is available online (
5.1.2. Parameter Setting
The experiments are implemented with MATLAB (version 9.11.01769968 (R2021b)) running on a 64-bit Windows with Intel (R) Xeon (R) E-2224 CPU 3.40GHz CPU, NVIDIA Quadro P1000, and 16.0 GB RAM.
ABC [88], PSO [89], CSO [90], DE [91], SSA [70], OFA [92], and SHADE [93] are used as benchmark algorithms for comparison. The number of trials for releasing a food resource of ABC is 20, the inertia weight of PSO is 0.4, the social factor of CSO is 0.1, the conjugate constant of DE is 0.9, the mutation factor of DE is 0.5, the historical memory size of SHADE is 100, the chaos disturbance factor parameter is 0.7, and the LF parameter is 1.5. The population in all algorithms is 100, and the number of runs is 30. Each algorithm repeats the experiment 30 times independently to obtain statistical results. The maximum number of function evaluations is 10,000. The solution schemes are tested on the CEC2020 function at 10D, 15D, and 20D.
5.1.3. Statistical Test
The significance level is used to compare whether the two algorithms significantly differ in performance. We use the Wilcoxon rank sum test with = 0.05 [94]. The original assumption is that the performance of TFSSA and the comparison algorithm is independent. When rejecting the original hypothesis, this paper uses three symbols to indicate whether there is a significant difference in the performance between TFSSA and the comparison algorithm.
(1) +: TFSSA performs significantly better than the comparison algorithm.
(2) =: The performance of TFSSA is not significantly related to the performance of the comparison algorithm.
(3) −: TFSSA’s performance is not significantly better than the comparison algorithm.
5.1.4. Solution Accuracy Analysis
This section displays the average value (Mean), standard deviation (Std), and Wilcoxon rank sum test results produced by various algorithms on CEC2020 for each test function. The best results from all experiments are highlighted in bold. The following is a complete description and analysis of the experimental results:
F1 in Table 2, Table 3 and Table 4 shows the optimization results of the unimodal function obtained by different algorithms. Under the circumstances of 10D, 15D, and 20D, the submitted TFSSA for a single-peak function has an average weight and standard difference capital compared to other methods.
F2–F4 in Table 2, Table 3 and Table 4 display the basic function improvement results obtained by the differential arithmetic method. Under the circumstances of 10D and 20D, TFSSA has the highest average CSO award among 10D, F4 has the highest average, and F2 and F3 have the highest functional average. For the standard difference direction, F2 and F3 shows the best results in 10D, and F3 shows the best results in 15D. Due to this and other arithmetic ratios, the proposed TFSSA has improved performance on basic functions. The number of events increased as a result of the above findings, and the supplied TFSSA displayed the ideal performance.
F5–F7 in Table 2, Table 3 and Table 4 show the results obtained by different algorithms for the optimization of mixed functions. It can be seen that in the case of 10D, TFSSA obtained the two best averages of F6 and F7; in the case of 15D, TFSSA obtained the best average of F6; in the case of 20D, TFSSA also obtained the best average of F6. SHADE obtains the best mean of F5 in 10D; CSO obtains the best mean of F5 and F7 in both 15D and 20D. This indicates that the stability of TFSSA slightly increases with the increase in the dimension of the solution decline. In particular, TFSSA has obvious advantages on the F6 function and achieves the best results in all dimensions compared to the comparison algorithm.
F8–F10 in Table 2, Table 3 and Table 4 show the results of the synthesis function optimization obtained by different algorithms. It can be seen that in the case of 10D, TFSSA obtained the best mean of F8; in the case of 15D, TFSSA obtained the best mean of F8 and F10; in the case of 20D, TFSSA also obtained the best mean of F9. In the case of 10D, the best means of F9 and F10 are obtained by SSA and DE; in the case of 15D, the best means of F8 and F9 are obtained by DE; in the case of 20D, the best means of F9 and F10 are obtained by PSO and SHADE.
In conclusion, TFSSA obtained seven, five, and six optimal averages and zero, two, and one suboptimal average among the ten functions in all dimensions, respectively, indicating that dimensional changes with less impact constrain the algorithm’s accuracy in finding solutions. In the case of 10D, SHADE obtains an optimal mean and three optimal standard deviations, SSA and DE each get an optimal mean, and CSO and DE each obtain an optimal standard deviation. Under the condition of 15D, the performance of the proposed TFSSA is challenged by CSO. CSO obtains three optimal means and one optimal standard deviation. At the same time, ABC and DE also achieved an optimal mean and standard deviation, and SHADE achieved three optimal means and the best standard deviation. In the 20D case, PSO and CSO obtain one and two best mean values, and PSO, DE, and SHADE get one, one, and four best standard deviations, respectively. According to NFL theory, it is almost impossible for one algorithm to solve all optimization problems efficiently. Therefore, the proposed TFSSA algorithm cannot obtain the best results on all classical test functions. However, compared with other algorithms, the best results it gets are still ideal, which verifies the superiority of the TFSSA algorithm to a certain extent. It can be seen from the results that the proposed TFSSA has good performance on CEC2020 at 10D, 15D, and 20D.
5.1.5. Algorithm Stability Analysis
From the results of the Wilcoxon test in Table 2, Table 3 and Table 4, it is observed that TFSSA significantly outperforms ABC, PSO, CSO, DE, and OFA on more than half of the functions of SHADE performance. Compared with the performance of DE and SHADE, at 15D, the performance of TFSSA is significantly improved on six functions, but the performance on four functions is reduced substantially. In other words, TFSSA performs much better than DE and SHADE at 15D. Compared with the performance of CSO, in the case of 10D, the performance of TFSSA is significantly improved on seven functions. Still, the performance of the two functions is reduced dramatically, and at 15D, the performance of TFSSA is considerably lower than that of CSO. Significantly improved performance on six functions but significantly reduced performance on two functions, with 20D TFSSA greatly enhanced performance on four parts but decreased performance considerably on six. It also shows that the stability of TFSSA at 10D is higher than that of 15D and higher than that of 20D to a certain extent.
5.1.6. Convergence Rate Analysis
This subsection presents the convergence rates obtained by different algorithms when solving the CEC2020 test function. The convergence speed of obtaining the optimal global solution is an important indicator for checking the performance of EA. Figure 3, Figure 4 and Figure 5 show TFSSA and comparison algorithms at 10D, 15D, and 20D on CEC2020, respectively, and the convergence plot obtained during the process of solving the test function. Among them, the abscissa represents the number of function evaluations, and the ordinate is the minimum value obtained each time the algorithm runs independently.
It can be seen that in the 10D case, on F1, F4, F6, F7, and F8, the convergence speed of TFSSA is significantly faster than most of the other comparison algorithms, among which F6 and F7 are more significant; in the 15D case, TFSSA converges considerably quicker than most of the different comparison algorithms on F1, F4, F5, F7, and F8; in the 20D case, TFSSA converges significantly faster than most of the other comparison algorithms on F1, F4, F5, F6, and F7, especially in the early stages of evolution of these classical test functions, showing faster convergence rates. In addition, the advantages of TFSSA on other parts are not obvious. Except for F2 and F3 at 10D, 15D, and 20D, F9 at 10D, 15D, and 20D, and F8 at 20D, the convergence speed of TFSSA is slightly worse than other algorithms, but the accuracy is still the best. The experiment shows that the TFSSA algorithm’s exploration ability in the later stage is relatively strong, which means that the proposed TFSSA algorithm can maintain relatively high population diversity and avoid premature convergence.
Overall, TFSSA showed the best convergence speed for most tested functions throughout the optimization process. Therefore, it can be concluded that the proposed TFSSA has a relatively good exploration ability on most of the test functions.
5.1.7. Sensitivity Analysis
This section investigates the sensitivity of TFSSA to (1) parameter a, (2) parameter , and (3) parameter c. This analysis helps determine which parameters are more robust and sensitive to various input values and which parameters have a greater impact on the accuracy of TFSSA. This study conducts a complete TFSSA design with some functions selected from the CEC2020 test suite. These functions include: CEC2020_F1 (20D), CEC2020_F2 (20D), CEC2020_F3 (20D), CEC2020_F1 (10D), CEC2020_F2 (10D), and CEC2020_F3 (10D). This experiment uses the same fitness function settings as before to ensure fairness. TFSSA’s sensitivity analysis results for the test function for all dimensions considered were investigated based on the average fitness value of 30 independent runs.
(1) Control parameter a: In the initialization of TFSSA, the control parameter a is involved. To check the sensitivity of TFSSA to a, different values of this parameter were simulated based on keeping other parameters unchanged, which are 0.75, 0.7, 0.65, and 0.6. The influence of different parameter values on the mean value of TFSSA is shown in Table 5.
(2) Control parameter : In the LF mechanism, control parameter is involved. To check the sensitivity of TFSSA to , different values of this parameter were simulated based on keeping other parameters unchanged, which are 1.4, 1.5, 1.6, and 1.7. The influence of different parameter values on the mean value of TFSSA is shown in Table 6.
(3) Control parameter c: In the adaptive hyperparameter, c is the adaptation factor of w, ensuring that w stays as a small value. To check the sensitivity of TFSSA to c, different values of this parameter were simulated based on keeping other parameters unchanged, which are 0.8, 0.85, 0.9, and 0.95. The influence of different parameter values on the mean value of TFSSA is shown in Table 7.
Overall, as can be seen from the results in Table 5, Table 6 and Table 7, TFSSA has a relatively good sensitivity to parameters a, , and c, providing relatively reasonable results. TFSSA produced the best results when a, , and c were 0.7, 1.5 and 0.9, respectively. It is of great help to study the sensitivity of the above control parameters to the performance of TFSSA. Furthermore, these parameters must be fine-tuned to help TFSSA obtain the best global solution.
5.1.8. Runtime Analysis
Table 8, Table 9 and Table 10 show the running time of TFSSA and the comparison algorithm at 10D, 15D, and 20D. It can be seen that the running time of TFSSA on all test functions is slightly longer than most comparison algorithms. The main reasons for the above phenomenon are as follows:
1. When mutating the optimal individual, TFSSA compares the calculated r with the random value and mutates the optimal individual. This stage is more expensive than the original SSA.
2. The optimal individual must reorder the fitness function values after passing through the -Tent chaotic mutation. Sorting is time-consuming, so this stage is also one of the main reasons for the increase in running time.
The running time of TFSSA in the study is slightly higher than most of these comparison algorithms. Still, in the end, considering the performance improvement, these additional running times are negligible to a certain extent.
5.2. Performance of Proposed Model
5.2.1. Description of Data
The utility and strength of our suggested strategy will be thoroughly investigated by selecting features from well-known datasets. Twenty-one datasets are from the UCI machine learning repository [95] and can be accessed online (
5.2.2. Parameter Configuration
Several top-of-the-line and most recent FS techniques are contrasted with the suggested approach, which is summarized as follows:
Genetic Algorithm (GA) [96].
Dragonfly Algorithm (DA) [97].
Ant Lion Optimizer (ALO) [98].
Sparrow Search Algorithm (SSA) [70].
Sine Cosine Algorithm (SCA) [99].
Particle Swarm Optimizer (PSO) [89].
binary Butterfly Optimization Algorithm (bBOA) [100].
Brain Storm Optimizer (BSO) [101].
Improved Sparrow Search Algorithm (ISSA) [102].
Grey Wolf Optimizer (GWO) [103].
Each algorithm runs 20 times independently with a random seed. For all subsequent tests, the maximum number of repetitions is set at 100. In the population, there are seven search agents. For our evaluations, we test our approach with a 10-fold cross-validation. Table 12 shows the global and algorithm-specific parameter settings. To ensure a fair comparison of the algorithms, the parameters of the algorithms are gathered from the literature. The main purpose of this research is to evaluate the performance of numerous FS methods compared to the proposed methodology. The K-NN classifier is a popular wrapper approach for FS. When K = 5, the method produces superior results.
5.2.3. Evaluation Criteria
For each experiment, we randomly split each dataset into three unequal parts at random: training, testing, and validation datasets, with a ratio of 6:2:2. The dataset partition process was repeated ten times in each 10-fold cross-validation, and the average performance of accuracy for these ten results is compared for all methods. The following assurances are captured from the validation data for each run:
Classification average accuracy (AvgPerf) is a metric that indicates how accurate the classifier is given the provided feature set. Equation (20) can be used to receive the classification average accuracy.
(20)
where M denotes the amount of times the optimizer is run to pick the feature subset, N denotes the number of points in the test set, denotes the output label of the classifier for data point denotes the data point i’s reference class label. If the two input labels are identical, the function returns 1 if they are. Otherwise, it returns 0.Statistical Best is the optimistic fitness value (the minimum value) obtained after each feature selection method runs M times, as shown in Equation (21).
(21)
where indicates the best result determined after i times of operation.Statistical Worst is the pessimistic result, which can be expressed as shown in Equation (22).
(22)
Statistical Mean is the average value of the solution obtained by running under the condition of M times, as shown in Equation (23).
(23)
Statistical Std is a representation of the variation in the obtained minimum (best) solutions for M different runs of a stochastic optimizer. Std is a stability and robustness metric for optimizers; if Std is small, the optimizer always converges to the same solution; on the contrary, the optimizer produces numerous random outcomes, as shown in Equation (24).
(24)
Selection average size (AVGSelectionSZ) represents the average amount of features selected, as shown in Equation (25).
(25)
where is the dimension of each dataset, and is the amount of on values for the vector x.Wilcoxon rank sum test is a nonparametric statistical test designed to see if the results of a proposed new technique are statistically different from those of other comparative techniques. The rank sum test produces a p-value parameter that compares the significance level of the two methods. The p-value is less than 0.05, which indicates that the two methods are significantly different [104,105].
5.2.4. Comparison of TFSSA and Other FS Methods
In this section, the performance of the best strategy, TFSSA, is compared to that of nine approaches (including the BSO, ALO, PSO, GWO, GA, bBOA, DA, SSA, and ISSA) that have been widely used to address the FS problem in the literature. Some performance indicators used to evaluate the algorithm’s performance include classification average accuracy, selected average feature number, selected average feature rate, statistical best fitness, statistical worst fitness, statistical mean fitness, statistical Std, calculation time, and Wilcoxon rank-sum test.
In Table 13, the classification average accuracy achieved by each algorithm is compared. TFSSA is preferred over other algorithms in most datasets except Exactly-1 and SonarMR. Furthermore, Figure 6 shows the overall average classification accuracy selected by different algorithms on all datasets. We can see that the proposed algorithm ranks first with a classification accuracy of 0.9011. This result confirms that the proposed TFSSA can effectively explore the solution search space and find the optimal feature subset with the highest classification accuracy.
Table 14 compares the average number and ratio of features selected by different algorithms. Both tables show that TFSSA outperforms the other algorithms in the 13 datasets. Although the number of features chosen by TFSSA is not optimal in the other eight datasets, it is not significantly different from other outperformed methods. Figure 7 shows the population’s average number of features and proportions chosen by the algorithm. The experiment shows that the average number of features and ratios selected by TFSSA in all datasets ranks first, with 30.97 and 0.468, respectively. Although the advantage is not apparent, it can prove that TFSSA outperforms other algorithms in most datasets to ensure high classification accuracy. In analyzing algorithm performance, we want to pay more attention to the classification average accuracy and the average number of features.
As a result, the number of selected attributes affected by the classification accuracy value is often slightly less than the fitness value. Table 15, Table 16, Table 17 and Table 18 present the statistical measures (best, worst, mean, and Std) obtained by different runs of the algorithm on each dataset. TFSSA has a lower fitness value than other algorithms, checking the results. Among them, the average fitness value of TFSSA maintains a leading edge in 17 datasets, and bBOA outperforms different algorithms in 5 datasets. The overall average fitness of TFSSA ranked first, with a value of 0.098. The best fitness value of TFSSA maintains the lead in all datasets except Exactly-1, and its overall best fitness value is 0.076, ranking first. The worst fitness value of TFSSA outperforms other algorithms in 17 datasets, bBOA outperforms different algorithms in 4 datasets, and GA outperforms other algorithms in dataset Clean-1. Table 18 shows that the standard deviation of TFSSA outperforms different algorithms in 21 datasets and Figure 8 compares the total average standard deviation for mean fitness values among algorithms, while the standard deviation of GA outperforms different algorithms in 8 datasets. The standard deviation of bBOA outperforms other algorithms in 2 datasets. Outperforming different algorithms, the standard deviation of GWO outperforms other algorithms in the dataset StatlogH.
The average execution time of each method in the experiment is shown in Table 19. Because almost all optimization algorithms employ the same amount of iterations, the computation time can be used to compare algorithm performance. We receive the following observations from Table 19. The ten EAs have intimate performances regarding the time consumption for all 21 datasets. As we all know, an EA-based feature selection technique requires a classifier to evaluate an individual. The time it takes the classifier to assess a set of features and samples is usually proportional to the number of features and samples. Therefore, the fitness function is the most time-consuming part of EA-based feature selection algorithms for datasets with many features or/and models, such as WaveformV2, Clean-2, and Semeion. The 10 EA-based algorithms used in the trials all had the same maximum number of evaluations as their termination conditions, which resulted in identical time consumption. Among them, TFSSA has the best computing time on seven datasets. In comparison, GWO performs better on five datasets, and GA performs better than other optimizers on six datasets, DA, SSA, and ISSA each have a better performance on one dataset.
In addition, Table 20 shows the Wilcoxon rank-sum test p-values at the 5% significance level for the Wilcoxon rank-sum test. A p-value of less than 0.05 implies that the null hypothesis of no meaningful difference at the 5% level is rejected. The p-values in Table 20 confirm that the results of TFSSA are significantly different from those of classical and top-of-the-line algorithms on most datasets. Specifically, in 12 datasets, the performance is outstanding, including BreastCWD, Clean-2, Exactly-1, Exactly-2, StatlogH, Lymphography, M-of-n, SonarMR, Spectheart, 3T Endgame, Vote, and Wine.
Overall, the results in Table 13, Table 14, Table 15, Table 16, Table 17 and Table 18 show that TFSSA can balance exploration and exploitation in the optimization search process. This experiment employed four large datasets: Clean-2 (No. 4), krvskpEW (No. 10), Penglung (No. 13), and Semeion (No. 14). The results indicate that TFSSA outperforms other algorithms in both small and large datasets. TFSSA outperforms all algorithms in classification average accuracy, selected average feature number, chose average feature rate, measures of fitness (best, worst, mean, and Std), and the Wilcoxon rank-sum test.
We can conclude from all of these experiments that employing the improved Tent chaos, LF strategy, and self-adaptive hyper-parameters improves the robustness and performance of the proposed algorithm. This method solves FS difficulties by combining global search algorithms (suitable for exploration) and local search algorithms (suitable for development). Establishing a balance between exploration and production in the FS problem is critical to avoiding many local solutions and discovering an accurate approximation of the optimal solution. This is the primary reason behind TFSSA’s improved performance compared to the comparative algorithm used in this study. TFSSA has the fewest features and the highest accuracy among the ten approaches. However, compared to the other methods utilized in this study, TFSSA takes more calculation time. Another drawback of the suggested random wrapper-based FS strategy is the imprecision with which the optimization results can be repeated. The algorithm’s subset of features selected for different applications has been noted, which may mislead users when determining which subset to evaluate.
6. Real-World Dataset Instances
COVID-19 is an infectious disease caused by SARS-CoV-2, which has led to an epidemic that has continued to this day and has become one of the epidemics with the most significant number of deaths in human history [106]. The first known patient with the disease was diagnosed in Wuhan, Hubei Province, China, at the end of 2019 (although the disease is likely to have infected humans before). Since then, the disease has been detected worldwide and is still spreading. At the same time, humanity hopes to defeat the virus through various technologies, so it has once again started a protracted war against the virus. According to research, Artificial Intelligence (AI) has become a weapon with great potential to fight SARS-CoV-2 [107].
This section employs the proposed TFSSA for 2019 Coronavirus Disease patient health prediction, as shown in Figure 9. The dataset of COVID-19 patients (
As can be seen from Figure 10, TFSSA achieves the highest average classification accuracy of 93.47% and the lowest average feature selection number of 2.1. On the other hand, the results reveal that for TFSSA inpatient health prediction, around three features were sufficient. According to the results, the most popular features were id, age, and nationality. The list of features selected by all FS algorithms is shown in Table 23, where the selected features are the main features selected by all FS algorithms in all experiments, and the features not shown in the table are the features that are eliminated. Furthermore, the data suggest that the TFSSA algorithm has never chosen symptom_4, symptom_5, or symptom_6. Further, to validate TFSSA’s classification performance, we try to remove symptoms 4, 5, and 6, and the difference is minor compared to previous experimental findings. As a result, these features cannot appropriately detect the data pattern in the patient health prediction process. The performance of TFSSA is observed after eliminating these characteristics, and the classification accuracy is barely affected. To continue studying the performance of TFSSA, we remove the original feature (ID) from the dataset. The experiment revealed that the classification average accuracy is about 91.3%. The researchers said that in the future, more abundant, detailed, and comprehensive clinical features should be collected to more accurately predict the health status of patients.
7. Conclusions
In this paper, we propose a TFSSA that mainly combines a Tent chaotic map, LF, and self-adaptive hyper-parameters to solve the optimization problems. First, we test the performance of TFSSA using the scientific standard test function—the CEC2020 benchmark function and compare it with seven methods in multiple aspects. Second, TFSSA combines a K-NN classifier to solve the FS problem in wrapper-based mode. Twenty-one datasets from the UC Irvine Machine Learning Repository are utilized to validate the proposed method’s performance. In addition, the method is also applied to the diagnosis and prediction of COVID-19. Nine criteria are reported to evaluate each technique: classification average accuracy, average selection size, average selection rate, measures of fitness (best, worst, mean, and Std), computation time, and rank-sum test. Comparing TFSSA with five top-of-the-line methods (BSO, ALO, PSO, GWO, and GA) and the four latest high-performance methods (bBOA, DA, SSA, and ISSA), the experimental results show TFSSA achieves the goal of lowering the number of features and boosting the model’s accuracy by removing as many irrelevant and redundant features as possible. Therefore, TFSSA can find the best feature subset and obtain high accuracy when applied to various FS tasks. During the experiments, we also found that multiple sophisticated initialization processes can be employed in TFSSA to improve the speed. How to strengthen multiple advanced initialization procedures will be our future work.
Conceptualization, Q.Y. and Y.G.; methodology, Q.Y. and Y.G.; software, Q.Y.; validation, Q.Y., Y.G. and Y.S.; formal analysis, Y.G. and Y.S.; investigation, Y.G.; resources, Y.G.; data curation, Q.Y. and Y.G.; writing—original draft preparation, Q.Y., Y.G. and Y.S.; writing—review and editing, Q.Y., Y.G. and Y.S.; visualization, Q.Y., Y.G. and Y.S.; supervision, Y.G. and Y.S.; project administration, Y.G.; funding acquisition, Y.G. All authors have read and agreed to the published version of the manuscript.
Datasets related to this article can be found at (
We acknowledge the valuable comments from the anonymous reviewers. We would also like to thank the Editors for their generous comments and support during the review process.
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 7. Comparison among algorithms’ total average number of features and the selected feature ratio.
Figure 8. Comparison of total average standard deviation for mean fitness values among algorithms.
CEC2020 test suite.
No. | Functions |
|
|
---|---|---|---|
Unimodal Function | 1 | CEC 2017 [ |
100 |
Basic |
2 | CEC 2014 [ |
1100 |
3 | CEC 2017 [ |
700 | |
4 | CEC 2017 [ |
1900 | |
Hybrid |
5 | CEC 2014 [ |
1700 |
6 | CEC 2017 [ |
1600 | |
7 | CEC 2014 [ |
2100 | |
Composition |
8 | CEC 2017 [ |
2200 |
9 | CEC 2017 [ |
2400 | |
10 | CEC 2017 [ |
2500 | |
Search Range = |
Mean, standard deviations, and Wilcoxon rank sum test results of different algorithms on CEC2020 at 10D.
ABC |
PSO |
CSO |
DE |
SSA |
OFA |
SHADE |
TFSSA |
|
---|---|---|---|---|---|---|---|---|
CEC2020_F1 | ||||||||
CEC2020_F2 | ||||||||
CEC2020_F3 | ||||||||
CEC2020_F4 | ||||||||
CEC2020_F5 | ||||||||
CEC2020_F6 | ||||||||
CEC2020_F7 | ||||||||
CEC2020_F8 | ||||||||
CEC2020_F9 | ||||||||
CEC2020_F10 | ||||||||
+/-/= | 6/0/4 | 5/4/1 | 7/2/1 | 6/2/2 | 5/0/5 | 4/3/3 | 6/2/2 |
Mean, standard deviations, and Wilcoxon rank sum test results of different algorithms on CEC2020 at 15D.
ABC |
PSO |
CSO |
DE |
SSA |
OFA |
SHADE |
TFSSA |
|
---|---|---|---|---|---|---|---|---|
CEC2020_F1 | ||||||||
CEC2020_F2 | ||||||||
CEC2020_F3 | ||||||||
CEC2020_F4 | ||||||||
CEC2020_F5 | ||||||||
CEC2020_F6 | ||||||||
CEC2020_F7 | ||||||||
CEC2020_F8 | ||||||||
CEC2020_F9 | ||||||||
CEC2020_F10 | ||||||||
+/-/= | 6/2/2 | 5/3/2 | 6/1/3 | 6/4/0 | 7/3/0 | 5/2/3 | 6/4/0 |
Mean, standard deviations, and Wilcoxon rank sum test results of different algorithms on CEC2020 at 20D.
ABC |
PSO |
CSO |
DE |
SSA |
OFA |
SHADE |
TFSSA |
|
---|---|---|---|---|---|---|---|---|
CEC2020_F1 | ||||||||
CEC2020_F2 | ||||||||
CEC2020_F3 | ||||||||
CEC2020_F4 | ||||||||
CEC2020_F5 | ||||||||
CEC2020_F6 | ||||||||
CEC2020_F7 | ||||||||
CEC2020_F8 | ||||||||
CEC2020_F9 | ||||||||
CEC2020_F10 | ||||||||
+/-/= | 5/4/1 | 5/3/2 | 4/6/0 | 5/3/2 | 4/4/2 | 5/3/2 | 6/4/0 |
The mean value of TFSSA under different values for parameter a.
a | CEC2020 Functions | |||||
---|---|---|---|---|---|---|
F1(20D) | F2(20D) | F3(20D) | F1(10D) | F2(10D) | F3(10D) | |
0.75 |
|
|
|
|
|
|
0.7 |
|
|
|
|
|
|
0.65 |
|
|
|
|
|
|
0.6 |
|
|
|
|
|
|
The mean value of TFSSA under different values for parameter
|
CEC2020 Functions | |||||
---|---|---|---|---|---|---|
F1(20D) | F2(20D) | F3(20D) | F1(10D) | F2(10D) | F3(10D) | |
1.4 |
|
|
|
|
|
|
1.5 |
|
|
|
|
|
|
1.6 |
|
|
|
|
|
|
1.7 |
|
|
|
|
|
|
The mean value of TFSSA under different values for parameter c.
c | CEC2020 Functions | |||||
---|---|---|---|---|---|---|
F1(20D) | F2(20D) | F3(20D) | F1(10D) | F2(10D) | F3(10D) | |
0.8 |
|
|
|
|
|
|
0.85 |
|
|
|
|
|
|
0.9 |
|
|
|
|
|
|
0.95 |
|
|
|
|
|
|
Running times of different algorithms on CEC2020 at 10D.
ABC | PSO | CSO | DE | SSA | OFA | SHADE | TFSSA |
---|---|---|---|---|---|---|---|
0.117 | 0.123 | 0.190 | 0.132 | 0.194 | 0.103 | 0.178 | 0.141 |
0.136 | 0.144 | 0.210 | 0.148 | 0.217 | 0.105 | 0.213 | 0.164 |
0.130 | 0.136 | 0.199 | 0.141 | 0.196 | 0.103 | 0.202 | 0.146 |
0.121 | 0.130 | 0.195 | 0.131 | 0.202 | 0.100 | 0.198 | 0.136 |
0.172 | 0.126 | 0.209 | 0.144 | 0.215 | 0.120 | 0.213 | 0.153 |
0.148 | 0.130 | 0.231 | 0.152 | 0.200 | 0.119 | 0.196 | 0.160 |
0.142 | 0.154 | 0.231 | 0.195 | 0.228 | 0.109 | 0.209 | 0.196 |
0.202 | 0.168 | 0.256 | 0.167 | 0.295 | 0.151 | 0.238 | 0.167 |
0.260 | 0.176 | 0.391 | 0.221 | 0.312 | 0.166 | 0.281 | 0.215 |
0.260 | 0.188 | 0.265 | 0.192 | 0.257 | 0.146 | 0.238 | 0.215 |
Running times of different algorithms on CEC2020 at 15D.
ABC | PSO | CSO | DE | SSA | OFA | SHADE | TFSSA |
---|---|---|---|---|---|---|---|
0.27314 | 0.17447 | 0.21609 | 0.18038 | 0.25067 | 0.13861 | 0.19847 | 0.19285 |
0.17174 | 0.14590 | 0.21438 | 0.23227 | 0.24845 | 0.12954 | 0.20751 | 0.19680 |
0.14075 | 0.14449 | 0.18067 | 0.17106 | 0.24346 | 0.12337 | 0.18830 | 0.19825 |
0.13625 | 0.13080 | 0.19684 | 0.15475 | 0.22128 | 0.12704 | 0.18636 | 0.17833 |
0.18806 | 0.15771 | 0.22111 | 0.19518 | 0.25376 | 0.12731 | 0.21287 | 0.20431 |
0.16081 | 0.16362 | 0.22160 | 0.18305 | 0.25280 | 0.11954 | 0.21252 | 0.19229 |
0.18113 | 0.16160 | 0.22597 | 0.18884 | 0.25475 | 0.13430 | 0.21125 | 0.20429 |
0.18804 | 0.17925 | 0.28799 | 0.19967 | 0.27466 | 0.15683 | 0.22740 | 0.23170 |
0.20090 | 0.22087 | 0.27973 | 0.26221 | 0.29791 | 0.17547 | 0.26660 | 0.25203 |
0.21197 | 0.21208 | 0.33228 | 0.21209 | 0.30600 | 0.18752 | 0.27462 | 0.23527 |
Running times of different algorithms on CEC2020 at 20D.
ABC | PSO | CSO | DE | SSA | OFA | SHADE | TFSSA |
---|---|---|---|---|---|---|---|
0.28840 | 0.19697 | 0.19697 | 0.17951 | 0.27007 | 0.12649 | 0.20115 | 0.22174 |
0.18208 | 0.22409 | 0.22409 | 0.28066 | 0.29550 | 0.15021 | 0.22463 | 0.22936 |
0.14329 | 0.19587 | 0.19587 | 0.19650 | 0.30428 | 0.12872 | 0.21098 | 0.22780 |
0.15119 | 0.20386 | 0.20386 | 0.19689 | 0.25890 | 0.12878 | 0.21079 | 0.20586 |
0.17717 | 0.22385 | 0.22385 | 0.20411 | 0.27980 | 0.14777 | 0.22679 | 0.23241 |
0.17056 | 0.20903 | 0.20903 | 0.19523 | 0.27751 | 0.14084 | 0.21278 | 0.22460 |
0.19126 | 0.22421 | 0.22421 | 0.20025 | 0.29574 | 0.16213 | 0.22949 | 0.21372 |
0.22325 | 0.27305 | 0.27305 | 0.24980 | 0.30859 | 0.19916 | 0.25927 | 0.25078 |
0.22689 | 0.32813 | 0.32813 | 0.29938 | 0.32780 | 0.22705 | 0.30924 | 0.28994 |
0.24966 | 0.31390 | 0.31390 | 0.25017 | 0.34546 | 0.18456 | 0.29214 | 0.26953 |
Dataset descriptions.
No. | Dataset | #Feat | #SMP | #CL | Area |
---|---|---|---|---|---|
1 | BreastCO | 9 | 699 | 2 | Medical |
2 | BreastCWD | 30 | 569 | 2 | Medical |
3 | Clean-1 | 166 | 476 | 2 | Physical |
4 | Clean-2 | 166 | 6598 | 2 | Physical |
5 | CongressVR | 16 | 435 | 2 | Social |
6 | Exactly-1 | 13 | 1000 | 2 | Biology |
7 | Exactly-2 | 13 | 1000 | 2 | Biology |
8 | StatlogH | 13 | 270 | 5 | Life |
9 | IonosphereVS | 34 | 351 | 2 | Physical |
10 | KrvskpEW | 36 | 3196 | 2 | Game |
11 | Lymphography | 18 | 148 | 4 | Medical |
12 | M-of-n | 13 | 1000 | 2 | Biology |
13 | Penglung | 325 | 73 | 2 | Biology |
14 | Semeion | 265 | 1593 | 2 | Computer |
15 | SonarMR | 60 | 208 | 2 | Physical |
16 | Spectheart | 22 | 267 | 2 | Life |
17 | 3T Endgame | 9 | 958 | 2 | Game |
18 | Vote | 16 | 300 | 2 | Life |
19 | WaveformV2 | 40 | 5000 | 3 | Physical |
20 | Wine | 13 | 178 | 3 | Physical |
21 | Zoology | 16 | 101 | 7 | Life |
Experiment parameter configuration.
Parameter Description | Value |
---|---|
a parameter in Tent chaos |
|
|
|
|
|
|
|
Count of runs (M) | 20 |
The amount of search agents | 7 |
The amount of T_max | 100 |
Problem Dimensions | No. of features in each datasets |
K for cross-validation | 10 |
Search field |
|
GA crossover ratio |
|
GA mutation ratio |
|
Selection strategy in GA | Roulette wheel |
A factors in WOA |
|
Acceleration factors in PSO |
|
Inertia index(w) in PSO |
|
A factors in GWO |
|
Mutation rate |
|
Parameter(a) in bBOA |
|
Parameter(c) in bBOA |
|
The amount of clusters in BSO | 5 |
Comparison of the classification accuracy of each algorithm.
No. | Datasets | ALO | BSO | GA | GWO | PSO | bBOA | DA | SSA | ISSA | TFSSA |
---|---|---|---|---|---|---|---|---|---|---|---|
1 | BreastCO | 0.9591 | 0.9200 | 0.9597 | 0.9603 | 0.9609 | 0.9286 | 0.9626 | 0.9600 | 0.9611 | 0.9668 |
2 | BreastCWD | 0.9392 | 0.9020 | 0.9488 | 0.9375 | 0.9385 | 0.9396 | 0.9385 | 0.9347 | 0.9396 | 0.9718 |
3 | Clean-1 | 0.8465 | 0.8261 | 0.8697 | 0.8580 | 0.8549 | 0.8562 | 0.8541 | 0.8431 | 0.8585 | 0.8923 |
4 | Clean-2 | 0.9496 | 0.9391 | 0.9423 | 0.9463 | 0.9465 | 0.9480 | 0.9487 | 0.9462 | 0.9510 | 0.9667 |
5 | CongressVR | 0.9370 | 0.8547 | 0.9413 | 0.9327 | 0.9235 | 0.9280 | 0.9318 | 0.9321 | 0.9349 | 0.9521 |
6 | Exactly-1 | 0.7061 | 0.6021 | 0.7306 | 0.7249 | 0.7471 | 0.8531 | 0.7481 | 0.7091 | 0.7197 | 0.8524 |
7 | Exactly-2 | 0.6980 | 0.6345 | 0.6940 | 0.6929 | 0.6959 | 0.6527 | 0.7007 | 0.6985 | 0.6977 | 0.7472 |
8 | StatlogH | 0.7773 | 0.6948 | 0.7867 | 0.7768 | 0.7788 | 0.7583 | 0.7773 | 0.7595 | 0.7842 | 0.8127 |
9 | IonosphereVS | 0.8595 | 0.8538 | 0.8938 | 0.8682 | 0.8485 | 0.8639 | 0.8708 | 0.8890 | 0.8826 | 0.9042 |
10 | KrvskpEW | 0.9006 | 0.7603 | 0.9215 | 0.9143 | 0.9200 | 0.8580 | 0.9269 | 0.8929 | 0.8980 | 0.9360 |
11 | Lymphography | 0.7863 | 0.6931 | 0.8164 | 0.7629 | 0.7906 | 0.8613 | 0.7793 | 0.7736 | 0.7880 | 0.8667 |
12 | M-of-n | 0.8184 | 0.7033 | 0.7988 | 0.8272 | 0.8425 | 0.8689 | 0.8293 | 0.8361 | 0.8549 | 0.9020 |
13 | Penglung | 0.8072 | 0.7676 | 0.6721 | 0.8341 | 0.8140 | 0.8482 | 0.8268 | 0.8331 | 0.7951 | 0.8745 |
14 | Semeion | 0.9584 | 0.9461 | 0.9557 | 0.9471 | 0.9476 | 0.9480 | 0.9521 | 0.9449 | 0.9504 | 0.9729 |
15 | SonarMR | 0.8487 | 0.7936 | 0.8750 | 0.8622 | 0.8667 | 0.8614 | 0.8506 | 0.8449 | 0.8506 | 0.8634 |
16 | Spectheart | 0.7881 | 0.7507 | 0.8097 | 0.7846 | 0.7841 | 0.7643 | 0.8000 | 0.7826 | 0.7871 | 0.8443 |
17 | 3T Endgame | 0.7587 | 0.6601 | 0.7609 | 0.7537 | 0.8622 | 0.8667 | 0.7564 | 0.7557 | 0.7546 | 0.8983 |
18 | Vote | 0.9258 | 0.8413 | 0.9333 | 0.9196 | 0.9258 | 0.9618 | 0.9227 | 0.9196 | 0.9200 | 0.9695 |
19 | WaveformV2 | 0.7066 | 0.6150 | 0.6921 | 0.7096 | 0.7192 | 0.7827 | 0.7154 | 0.7091 | 0.7044 | 0.7929 |
20 | Wine | 0.9543 | 0.8652 | 0.9536 | 0.9476 | 0.9521 | 0.9474 | 0.9551 | 0.9506 | 0.9566 | 0.9843 |
21 | Zoology | 0.9216 | 0.8131 | 0.9294 | 0.9525 | 0.9451 | 0.8827 | 0.9359 | 0.9476 | 0.9307 | 0.9525 |
AVG. | 0.8499 | 0.7827 | 0.8517 | 0.8530 | 0.8602 | 0.8657 | 0.8563 | 0.8506 | 0.8533 | 0.9011 |
The bolded values represent the best outcomes.
Comparison of the selected average No. of features (AVG.NOF.) and the selected feature ratio (AVG_Ri) of each algorithm.
No. | Dataset | ALO | BSO | GA | GWO | PSO | bBOA | DA | SSA | ISSA | TFSSA |
AVG.NOF. | AVG.NOF. | AVG.NOF. | AVG.NOF. | AVG.NOF. | AVG.NOF. | AVG.NOF. | AVG.NOF. | AVG.NOF. | AVG.NOF. | ||
(AVG_Ri.) | (AVG_Ri.) | (AVG_Ri.) | (AVG_Ri.) | (AVG_Ri.) | (AVG_Ri.) | (AVG_Ri.) | (AVG_Ri.) | (AVG_Ri.) | (AVG_Ri.) | ||
1 | BreastCO | 7.00 |
6.40 |
6.10 |
6.90 |
5.70 |
5.60 |
6.27 |
7.20 |
5.70 |
4.40
|
2 | BreastCWD | 24.27 |
13.73 |
12.20 |
19.00 |
18.33 |
16.80 |
20.00 |
18.27 |
20.47 |
8.40
|
3 | Clean-1 | 132.00 |
98.73 |
98.90 |
109.60 |
104.93 |
91.80 |
109.67 |
94.87 |
90.20
|
90.27 |
4 | Clean-2 | 95.00 |
101.00 |
94.10 |
106.00 |
109.40 |
92.40 |
100.40 |
90.40 |
92.40 |
90.28
|
5 | CongressVR | 9.87 |
7.53 |
7.10 |
9.80 |
10.80 |
6.40 |
10.87 |
8.40 |
9.00 |
6.15
|
6 | Exactly-1 | 12.87 |
7.73 |
8.10 |
12.07 |
9.00 |
7.60 |
10.53 |
12.80 |
10.47 |
6.48
|
7 | Exactly-2 | 8.40 |
6.27 |
7.10 |
7.53 |
9.40 |
4.80 |
8.67 |
6.27 |
9.00 |
4.62
|
8 | StatlogH | 10.40 |
6.60 |
6.60 |
8.80 |
9.07 |
5.80 |
9.60 |
7.47 |
8.47 |
4.86
|
9 | IonosphereVS | 20.13 |
15.93 |
13.50
|
17.33 |
19.20 |
16.20 |
18.00 |
19.67 |
19.07 |
17.14 |
10 | KrvskpEW | 35.80 |
17.80 |
18.00 |
31.60 |
25.60 |
17.60 |
28.60 |
29.40 |
20.80 |
16.91
|
11 | Lymphography | 13.33 |
9.47 |
8.90 |
11.80 |
11.73 |
8.40
|
12.53 |
12.20 |
8.87 |
9.17 |
12 | M-of-n | 11.27 |
6.90 |
7.68 |
11.27 |
10.87 |
6.80 |
12.13 |
12.33 |
10.67 |
6.30
|
13 | Penglung | 172.07 |
160.60 |
153.00
|
162.80 |
183.33 |
172.00 |
175.20 |
162.33 |
182.67 |
161.42 |
14 | Semeion | 187.80 |
162.00 |
149.40 |
203.60 |
171.60 |
143.20 |
193.00 |
161.80 |
194.40 |
142.38
|
15 | SonarMR | 48.00 |
30.60 |
30.30 |
41.60 |
37.60 |
32.80 |
29.40 |
34.13 |
37.13 |
22.36
|
16 | Spectheart | 13.87 |
10.87 |
7.00
|
13.20 |
12.07 |
10.80 |
14.67 |
11.33 |
9.60 |
8.60 |
17 | 3T Endgame | 8.80 |
5.88 |
5.80 |
7.53 |
6.73 |
5.60 |
7.20 |
8.07 |
7.47 |
5.29
|
18 | Vote | 8.40 |
7.87 |
5.80 |
8.47 |
9.33 |
5.20
|
8.87 |
8.53 |
9.60 |
8.67 |
19 | WaveformV2 | 39.60 |
29.00 |
30.40 |
36.60 |
35.80 |
25.00 |
36.00 |
37.20 |
34.40 |
24.56
|
20 | Wine | 11.07 |
6.67 |
6.73 |
10.73 |
10.07 |
6.20
|
9.53 |
9.07 |
9.40 |
6.34 |
21 | Zoology | 11.67 |
7.67 |
5.35 |
12.40 |
11.80 |
5.20
|
11.47 |
11.93 |
9.60 |
5.78 |
AVG. | 41.98 |
34.25 |
32.48 |
40.41 |
39.16 |
32.68 |
39.65 |
36.37 |
38.07 |
30.97 |
The bolded values represent the best outcomes.
Comparison of the average fitness measure of each algorithm.
No. | Dataset | ALO | BSO | GA | GWO | PSO | bBOA | DA | SSA | ISSA | TFSSA |
---|---|---|---|---|---|---|---|---|---|---|---|
1 | BreastCO | 0.048 | 0.084 | 0.046 | 0.047 | 0.045 | 0.040 | 0.041 | 0.046 | 0.048 | 0.032 |
2 | BreastCWD | 0.068 | 0.102 | 0.055 | 0.068 | 0.068 | 0.042 | 0.068 | 0.067 | 0.071 | 0.045 |
3 | Clean-1 | 0.160 | 0.177 | 0.134 | 0.147 | 0.150 | 0.113 | 0.151 | 0.147 | 0.160 | 0.108 |
4 | Clean-2 | 0.055 | 0.065 | 0.062 | 0.060 | 0.060 | 0.051 | 0.057 | 0.054 | 0.058 | 0.041 |
5 | CongressVR | 0.069 | 0.149 | 0.063 | 0.073 | 0.082 | 0.045 | 0.074 | 0.042 | 0.072 | 0.035 |
6 | Exactly-1 | 0.301 | 0.400 | 0.270 | 0.282 | 0.257 | 0.040 | 0.257 | 0.286 | 0.298 | 0.229 |
7 | Exactly-2 | 0.305 | 0.367 | 0.308 | 0.310 | 0.308 | 0.260 | 0.303 | 0.306 | 0.303 | 0.240 |
8 | StatlogH | 0.228 | 0.307 | 0.216 | 0.228 | 0.226 | 0.180 | 0.228 | 0.220 | 0.244 | 0.185 |
9 | IonosphereVS | 0.145 | 0.149 | 0.109 | 0.136 | 0.124 | 0.096 | 0.133 | 0.122 | 0.116 | 0.081 |
10 | KrvskpEW | 0.108 | 0.242 | 0.083 | 0.094 | 0.086 | 0.054 | 0.080 | 0.140 | 0.116 | 0.044 |
11 | Lymphography | 0.219 | 0.309 | 0.187 | 0.241 | 0.214 | 0.189 | 0.225 | 0.216 | 0.231 | 0.109 |
12 | M-of-n | 0.188 | 0.299 | 0.205 | 0.180 | 0.164 | 0.027 | 0.178 | 0.152 | 0.172 | 0.024 |
13 | Penglung | 0.196 | 0.235 | 0.129 | 0.169 | 0.190 | 0.118 | 0.177 | 0.209 | 0.170 | 0.106 |
14 | Semeion | 0.045 | 0.049 | 0.039 | 0.050 | 0.059 | 0.036 | 0.055 | 0.057 | 0.052 | 0.021 |
15 | SonarMR | 0.158 | 0.209 | 0.128 | 0.143 | 0.138 | 0.086 | 0.155 | 0.154 | 0.159 | 0.079 |
16 | Spectheart | 0.216 | 0.252 | 0.192 | 0.219 | 0.219 | 0.160 | 0.205 | 0.217 | 0.220 | 0.120 |
17 | 3T Endgame | 0.249 | 0.342 | 0.243 | 0.252 | 0.253 | 0.205 | 0.249 | 0.251 | 0.251 | 0.219 |
18 | Vote | 0.079 | 0.162 | 0.070 | 0.085 | 0.079 | 0.044 | 0.082 | 0.085 | 0.085 | 0.037 |
19 | WaveformV2 | 0.300 | 0.386 | 0.319 | 0.297 | 0.287 | 0.265 | 0.291 | 0.301 | 0.298 | 0.254 |
20 | Wine | 0.054 | 0.139 | 0.051 | 0.060 | 0.055 | 0.023 | 0.052 | 0.050 | 0.056 | 0.023 |
21 | Zoology | 0.085 | 0.190 | 0.073 | 0.055 | 0.062 | 0.034 | 0.071 | 0.075 | 0.059 | 0.021 |
AVG. | 0.156 | 0.220 | 0.142 | 0.152 | 0.149 | 0.100 | 0.149 | 0.152 | 0.154 | 0.098 |
The bolded values represent the best outcomes.
Comparison of the best fitness measure of each algorithm.
No. | Dataset | ALO | BS0 | GA | GWO | PSO | bBOA | DA | SSA | ISSA | TFSSA |
---|---|---|---|---|---|---|---|---|---|---|---|
1 | BreastCO | 0.038 | 0.046 | 0.040 | 0.038 | 0.039 | 0.024 | 0.031 | 0.038 | 0.038 | 0.022 |
2 | BreastCWD | 0.066 | 0.065 | 0.048 | 0.048 | 0.049 | 0.032 | 0.051 | 0.052 | 0.059 | 0.029 |
3 | Clean-1 | 0.118 | 0.130 | 0.122 | 0.117 | 0.100 | 0.088 | 0.118 | 0.100 | 0.122 | 0.074 |
4 | Clean-2 | 0.049 | 0.058 | 0.062 | 0.056 | 0.058 | 0.037 | 0.050 | 0.052 | 0.054 | 0.033 |
5 | CongressVR | 0.044 | 0.076 | 0.054 | 0.042 | 0.048 | 0.030 | 0.035 | 0.045 | 0.041 | 0.026 |
6 | Exactly-1 | 0.267 | 0.328 | 0.015 | 0.173 | 0.138 | 0.005 | 0.155 | 0.089 | 0.229 | 0.224 |
7 | Exactly-2 | 0.252 | 0.296 | 0.295 | 0.279 | 0.275 | 0.225 | 0.238 | 0.237 | 0.270 | 0.221 |
8 | StatlogH | 0.172 | 0.206 | 0.202 | 0.189 | 0.178 | 0.138 | 0.159 | 0.163 | 0.194 | 0.134 |
9 | IonosphereVS | 0.111 | 0.101 | 0.099 | 0.088 | 0.081 | 0.060 | 0.104 | 0.092 | 0.078 | 0.056 |
10 | KruskpLW | 0.093 | 0.133 | 0.063 | 0.090 | 0.052 | 0.036 | 0.062 | 0.084 | 0.111 | 0.032 |
11 | Lymphography | 0.165 | 0.220 | 0.168 | 0.193 | 0.179 | 0.183 | 0.166 | 0.168 | 0.169 | 0.064 |
12 | M-of-n | 0.160 | 0.170 | 0.140 | 0.128 | 0.064 | 0.005 | 0.157 | 0.101 | 0.035 | 0.003 |
13 | Penglung | 0.085 | 0.085 | 0.137 | 0.085 | 0.086 | 0.033 | 0.035 | 0.062 | 0.112 | 0.029 |
14 | Semeion | 0.041 | 0.046 | 0.033 | 0.044 | 0.042 | 0.029 | 0.040 | 0.047 | 0.045 | 0.020 |
15 | SonarMR | 0.128 | 0.139 | 0.109 | 0.090 | 0.091 | 0.072 | 0.113 | 0.081 | 0.129 | 0.069 |
16 | Spectheart | 0.144 | 0.198 | 0.170 | 0.149 | 0.166 | 0.122 | 0.142 | 0.159 | 0.173 | 0.118 |
17 | 3T Endgame | 0.213 | 0.252 | 0.232 | 0.223 | 0.204 | 0.195 | 0.217 | 0.219 | 0.213 | 0.183 |
18 | Vote | 0.043 | 0.065 | 0.061 | 0.051 | 0.039 | 0.016 | 0.051 | 0.060 | 0.050 | 0.012 |
19 | WaveformLW | 0.294 | 0.338 | 0.312 | 0.283 | 0.271 | 0.254 | 0.278 | 0.291 | 0.291 | 0.250 |
20 | Wine | 0.029 | 0.061 | 0.038 | 0.019 | 0.028 | 0.005 | 0.031 | 0.028 | 0.016 | 0.003 |
21 | Zoology | 0.026 | 0.025 | 0.061 | 0.007 | 0.008 | 0.002 | 0.007 | 0.009 | 0.026 | 0.002 |
AVG. | 0.121 | 0.145 | 0.117 | 0.114 | 0.105 | 0.076 | 0.107 | 0.104 | 0.117 | 0.076 |
The bolded values represent the best outcomes.
Comparison of the worst fitness measure of each algorithm.
No. | Dataset | ALO | BSO | GA | GWO | PSO | bBOA | DA | SSA | ISSA | TFSSA |
---|---|---|---|---|---|---|---|---|---|---|---|
1 | BreastCO | 0.059 | 0.196 | 0.051 | 0.054 | 0.06 | 0.041 | 0.059 | 0.056 | 0.058 | 0.036 |
2 | BreastCWD | 0.083 | 0.144 | 0.063 | 0.085 | 0.078 | 0.049 | 0.09 | 0.095 | 0.088 | 0.049 |
3 | Clean-1 | 0.193 | 0.208 | 0.143 | 0.187 | 0.186 | 0.138 | 0.178 | 0.214 | 0.200 | 0.153 |
4 | Clean-2 | 0.060 | 0.073 | 0.071 | 0.063 | 0.061 | 0.068 | 0.059 | 0.057 | 0.064 | 0.043 |
5 | CongressVR | 0.110 | 0.267 | 0.083 | 0.110 | 0.149 | 0.058 | 0.107 | 0.096 | 0.120 | 0.053 |
6 | Exactly-1 | 0.343 | 0.448 | 0.378 | 0.344 | 0.384 | 0.115 | 0.319 | 0.375 | 0.335 | 0.285 |
7 | Exactly-2 | 0.355 | 0.517 | 0.331 | 0.333 | 0.335 | 0.291 | 0.330 | 0.337 | 0.363 | 0.287 |
8 | StatlogH | 0.289 | 0.378 | 0.261 | 0.256 | 0.288 | 0.195 | 0.284 | 0.277 | 0.299 | 0.191 |
9 | IonosphereVS | 0.168 | 0.195 | 0.134 | 0.179 | 0.163 | 0.118 | 0.157 | 0.155 | 0.157 | 0.114 |
10 | KruskpLW | 0.118 | 0.344 | 0.150 | 0.096 | 0.164 | 0.064 | 0.097 | 0.176 | 0.121 | 0.060 |
11 | Lymphography | 0.251 | 0.378 | 0.220 | 0.299 | 0.276 | 0.194 | 0.303 | 0.261 | 0.299 | 0.146 |
12 | M-of-n | 0.224 | 0.391 | 0.288 | 0.235 | 0.287 | 0.110 | 0.210 | 0.236 | 0.212 | 0.206 |
13 | Penglung | 0.300 | 0.460 | 0.190 | 0.246 | 0.328 | 0.169 | 0.326 | 0.379 | 0.273 | 0.153 |
14 | Semeion | 0.049 | 0.056 | 0.043 | 0.064 | 0.072 | 0.049 | 0.070 | 0.077 | 0.065 | 0.025 |
15 | SonarMR | 0.216 | 0.253 | 0.156 | 0.218 | 0.187 | 0.109 | 0.217 | 0.198 | 0.214 | 0.103 |
16 | Spectheart | 0.271 | 0.322 | 0.218 | 0.265 | 0.265 | 0.209 | 0.252 | 0.271 | 0.262 | 0.201 |
17 | 3T Endgame | 0.275 | 0.436 | 0.255 | 0.309 | 0.331 | 0.216 | 0.293 | 0.307 | 0.293 | 0.225 |
18 | Vote | 0.113 | 0.256 | 0.088 | 0.169 | 0.119 | 0.057 | 0.124 | 0.138 | 0.118 | 0.056 |
19 | WaveformV2 | 0.304 | 0.434 | 0.319 | 0.316 | 0.303 | 0.265 | 0.299 | 0.313 | 0.305 | 0.259 |
20 | Wine | 0.075 | 0.303 | 0.082 | 0.142 | 0.075 | 0.028 | 0.086 | 0.077 | 0.076 | 0.026 |
21 | Zoology | 0.158 | 0.430 | 0.101 | 0.203 | 0.182 | 0.048 | 0.125 | 0.181 | 0.107 | 0.039 |
AVG. | 0.191 | 0.309 | 0.173 | 0.199 | 0.204 | 0.123 | 0.190 | 0.204 | 0.192 | 0.129 |
The bolded values represent the best outcomes.
Comparison of the standard deviation fitness measure of each algorithm.
No. | Dataset | ALO | BSO | GA | GWO | PSO | bBOA | DA | SSA | ISSA | TFSSA |
---|---|---|---|---|---|---|---|---|---|---|---|
1 | BreastCO | 0.008 | 0.023 | 0.005 | 0.011 | 0.008 | 0.006 | 0.011 | 0.013 | 0.009 | 0.005 |
2 | BreastCWD | 0.007 | 0.044 | 0.003 | 0.006 | 0.006 | 0.003 | 0.010 | 0.005 | 0.005 | 0.003 |
3 | Clean-1 | 0.017 | 0.048 | 0.008 | 0.019 | 0.026 | 0.010 | 0.021 | 0.016 | 0.020 | 0.018 |
4 | Clean-2 | 0.023 | 0.036 | 0.136 | 0.051 | 0.067 | 0.012 | 0.038 | 0.066 | 0.025 | 0.010 |
5 | CongressVR | 0.030 | 0.066 | 0.014 | 0.013 | 0.019 | 0.020 | 0.022 | 0.023 | 0.025 | 0.011 |
6 | Exactly-1 | 0.031 | 0.051 | 0.020 | 0.023 | 0.029 | 0.011 | 0.028 | 0.035 | 0.028 | 0.020 |
7 | Exactly-2 | 0.019 | 0.023 | 0.012 | 0.022 | 0.023 | 0.059 | 0.015 | 0.019 | 0.021 | 0.054 |
8 | StatlogH | 0.009 | 0.094 | 0.033 | 0.002 | 0.045 | 0.008 | 0.014 | 0.037 | 0.005 | 0.011 |
9 | IonosphereVS | 0.026 | 0.049 | 0.016 | 0.030 | 0.028 | 0.014 | 0.041 | 0.032 | 0.042 | 0.010 |
10 | KrvskpEW | 0.020 | 0.074 | 0.054 | 0.030 | 0.058 | 0.033 | 0.018 | 0.035 | 0.049 | 0.031 |
11 | Lymphography | 0.025 | 0.037 | 0.013 | 0.040 | 0.029 | 0.018 | 0.023 | 0.030 | 0.021 | 0.018 |
12 | M-of-n | 0.035 | 0.036 | 0.016 | 0.031 | 0.027 | 0.035 | 0.030 | 0.031 | 0.025 | 0.015 |
13 | Penglung | 0.020 | 0.053 | 0.006 | 0.026 | 0.034 | 0.007 | 0.023 | 0.025 | 0.021 | 0.018 |
14 | Semeion | 0.019 | 0.057 | 0.009 | 0.029 | 0.019 | 0.010 | 0.019 | 0.024 | 0.020 | 0.008 |
15 | SonarMR | 0.004 | 0.088 | 0.003 | 0.012 | 0.012 | 0.001 | 0.009 | 0.009 | 0.006 | 0.001 |
16 | Spectheart | 0.012 | 0.067 | 0.011 | 0.033 | 0.012 | 0.010 | 0.013 | 0.014 | 0.015 | 0.012 |
17 | 3T Endgame | 0.035 | 0.128 | 0.015 | 0.047 | 0.050 | 0.044 | 0.037 | 0.047 | 0.028 | 0.041 |
18 | Vote | 0.021 | 0.026 | 0.009 | 0.022 | 0.028 | 0.016 | 0.018 | 0.029 | 0.022 | 0.015 |
19 | WaveformV2 | 0.004 | 0.005 | 0.004 | 0.002 | 0.001 | 0.001 | 0.003 | 0.002 | 0.004 | 0.001 |
20 | Wine | 0.072 | 0.102 | 0.018 | 0.046 | 0.077 | 0.056 | 0.077 | 0.093 | 0.052 | 0.018 |
21 | Zoology | 0.007 | 0.002 | 0.001 | 0.005 | 0.002 | 0.004 | 0.003 | 0.006 | 0.004 | 0.002 |
The bolded values represent the best outcomes.
Comparison of the running time of each algorithm.
No. | Dataset | ALO | BSO | GA | GWO | PSO | bBOA | DA | SSA | ISSA | TFSSA |
---|---|---|---|---|---|---|---|---|---|---|---|
1 | BreastCO | 4.90 | 3.02 | 2.34 | 3.45 | 2.39 | 2.48 | 2.36 | 2.32 | 2.45 | 2.27 |
2 | BreastCWD | 2.87 | 2.85 | 2.37 | 3.61 | 2.43 | 2.82 | 2.41 | 2.36 | 2.35 | 2.32 |
3 | Clean-1 | 5.31 | 3.58 | 4.07 | 3.39 | 3.66 | 4.05 | 3.61 | 3.50 | 3.54 | 3.84 |
4 | Clean-2 | 310.83 | 223.69 | 279.32 | 158.67 | 227.16 | 171.84 | 223.70 | 173.07 | 182.94 | 159.32 |
5 | CongressVR | 2.88 | 3.33 | 2.81 | 3.32 | 2.64 | 3.03 | 2.59 | 2.61 | 2.72 | 2.59 |
6 | Exactly-1 | 3.92 | 4.58 | 3.85 | 4.04 | 4.53 | 4.06 | 4.63 | 5.18 | 4.65 | 4.96 |
7 | Exactly-2 | 4.22 | 4.62 | 4.15 | 4.52 | 4.82 | 4.60 | 4.88 | 4.20 | 4.22 | 4.20 |
8 | StatlogH | 2.69 | 2.96 | 2.46 | 3.09 | 2.49 | 2.78 | 2.52 | 2.62 | 2.47 | 2.50 |
9 | IonosphereVS | 3.14 | 3.10 | 2.58 | 3.25 | 2.64 | 2.96 | 2.60 | 2.54 | 2.57 | 2.47 |
10 | KrvskpEW | 18.16 | 11.56 | 10.42 | 9.53 | 17.16 | 13.84 | 15.89 | 13.89 | 13.03 | 12.07 |
11 | Lymphography | 2.68 | 2.94 | 2.46 | 2.98 | 3.04 | 2.82 | 2.38 | 2.87 | 2.91 | 2.69 |
12 | M-of-n | 4.08 | 4.07 | 3.53 | 3.39 | 4.56 | 4.04 | 3.74 | 4.26 | 4.14 | 4.19 |
13 | Penglung | 7.65 | 3.10 | 2.51 | 2.49 | 2.56 | 4.13 | 2.55 | 2.50 | 2.50 | 2.45 |
14 | Semeion | 28.41 | 14.33 | 13.10 | 31.67 | 24.51 | 19.92 | 24.06 | 21.82 | 19.21 | 15.45 |
15 | SonarMR | 3.30 | 2.93 | 2.39 | 2.97 | 2.62 | 2.92 | 2.72 | 2.75 | 2.59 | 2.45 |
16 | Spectheart | 2.88 | 2.96 | 2.45 | 3.00 | 2.40 | 2.80 | 2.38 | 2.40 | 2.38 | 2.29 |
17 | 3T Endgame | 4.36 | 3.99 | 3.28 | 4.38 | 4.49 | 3.91 | 4.38 | 4.25 | 4.10 | 4.45 |
18 | Vote | 2.89 | 3.26 | 2.62 | 3.25 | 2.57 | 2.82 | 2.60 | 2.53 | 2.62 | 2.47 |
19 | WaveformV2 | 40.51 | 25.03 | 23.48 | 20.63 | 27.09 | 35.56 | 43.72 | 34.14 | 36.64 | 21.26 |
20 | Wine | 2.68 | 2.92 | 2.46 | 3.13 | 2.45 | 2.68 | 2.43 | 2.43 | 2.47 | 2.52 |
21 | Zoology | 2.79 | 4.85 | 2.33 | 3.25 | 2.24 | 2.66 | 2.30 | 2.21 | 2.19 | 2.15 |
p-values of the Wilcoxon test of TFSSA vs. others.
Dataset | ALO | BSO | GA | GWO | PSO | bBOA | DA | SSA | ISSA | TFSSA | |
---|---|---|---|---|---|---|---|---|---|---|---|
1 | BreastCO |
|
|
|
|
|
|
|
|
|
|
2 | BreastCWD |
|
|
|
|
|
|
|
|
|
|
3 | Clean-1 |
|
|
|
|
|
|
|
|
|
|
4 | Clean-2 |
|
|
|
|
|
|
|
|
|
|
5 | CongressVR |
|
|
|
|
|
|
|
|
|
|
6 | Exactly-1 |
|
|
|
|
|
|
|
|
|
|
7 | Exactly-2 |
|
|
|
|
|
|
|
|
|
|
8 | StatlogH |
|
|
|
|
|
|
|
|
|
|
9 | IonosphereVS |
|
|
|
|
|
|
|
|
|
|
10 | KrvskpEW |
|
|
|
|
|
|
|
|
|
|
11 | Lymphography |
|
|
|
|
|
|
|
|
|
|
12 | M-of-n |
|
|
|
|
|
|
|
|
|
|
13 | Penglung |
|
|
|
|
|
|
|
|
|
|
14 | semeion |
|
|
|
|
|
|
|
|
|
|
15 | SonarMR |
|
|
|
|
|
|
|
|
|
|
16 | Spectheart |
|
|
|
|
|
|
|
|
|
|
17 | 3T Endgame |
|
|
|
|
|
|
|
|
|
|
18 | Vote |
|
|
|
|
|
|
|
|
|
|
19 | WaveformV2 |
|
|
|
|
|
|
|
|
|
|
20 | Wine |
|
|
|
|
|
|
|
|
|
|
21 | Zoology |
|
|
|
|
|
|
|
|
|
|
The p ≥ 0.05 are underlined.
COVID-19 dataset description.
Dateset | No. Features | No. Instances | Area |
---|---|---|---|
COVID-19 | 15 | 1085 | Medical |
The description of the 2019 Coronavirus Disease dataset.
No. | Features | Feature Description |
---|---|---|
1 | code(id) | Patients’ identification numbers |
2 | location | The place where patients are situated |
3 | nationality | The country from which the patients come |
4 | gender | The patients’ gender |
5 | age | How old patients are |
6 | sym_on | When people first show symptoms |
7 | hosp_vis | The date patients visit hospital |
8 | vis_wuhan | Whether or not the patients visited Wuhan, CN |
9 | from_wuhan | Whether or not the patients from Wuhan, CN |
10 | symptom_1 | One of the symptoms encountered by patients |
11 | symptom_2 | One of the symptoms encountered by patients |
12 | symptom_3 | One of the symptoms encountered by patients |
13 | symptom_4 | One of the symptoms encountered by patients |
14 | symptom_5 | One of the symptoms encountered by patients |
15 | symptom_6 | One of the symptoms encountered by patients |
The list of features selected by all FS algorithms.
Algorithm | id | Age | Nationality | sym_on | from_wuhan |
---|---|---|---|---|---|
ALO | ✓ | ✓ | ✓ | ||
BSO | ✓ | ✓ | ✓ | ||
GA | ✓ | ✓ | ✓ | ||
GWO | ✓ | ✓ | ✓ | ||
PSO | ✓ | ✓ | ✓ | ||
bBOA | ✓ | ✓ | |||
DA | ✓ | ✓ | ✓ | ✓ | |
SSA | ✓ | ✓ | ✓ | ✓ | |
ISSA | ✓ | ✓ | ✓ | ✓ | |
TFSSA | ✓ | ✓ |
References
1. Too, J.; Mirjalili, S. A hyper learning binary dragonfly algorithm for feature selection: A COVID-19 case study. Knowl.-Based Syst.; 2021; 212, 106553. [DOI: https://dx.doi.org/10.1016/j.knosys.2020.106553]
2. Frawley, W.J.; Piatetsky-Shapiro, G.; Matheus, C.J. Knowledge discovery in databases: An overview. AI Mag.; 1992; 13, 57.
3. Cios, K.J.; Pedrycz, W.; Swiniarski, R.W. Data mining and knowledge discovery. Data Mining Methods for Knowledge Discovery; Springer: Berlin/Heidelberg, Germany, 1998; pp. 1-26.
4. Gandomi, A.H.; Alavi, A.H. Krill herd: A new bio-inspired optimization algorithm. Commun. Nonlinear Sci. Numer. Simul.; 2012; 17, pp. 4831-4845. [DOI: https://dx.doi.org/10.1016/j.cnsns.2012.05.010]
5. García, S.; Ramírez-Gallego, S.; Luengo, J.; Benítez, J.M.; Herrera, F. Big data preprocessing: Methods and prospects. Big Data Anal.; 2016; 1, 9. [DOI: https://dx.doi.org/10.1186/s41044-016-0014-0]
6. Alasadi, S.A.; Bhaya, W.S. Review of data preprocessing techniques in data mining. J. Eng. Appl. Sci.; 2017; 12, pp. 4102-4107.
7. Mishra, P.; Biancolillo, A.; Roger, J.M.; Marini, F.; Rutledge, D.N. New data preprocessing trends based on ensemble of multiple preprocessing techniques. TrAC Trends Anal. Chem.; 2020; 132, 116045. [DOI: https://dx.doi.org/10.1016/j.trac.2020.116045]
8. Kamiran, F.; Calders, T. Data preprocessing techniques for classification without discrimination. Knowl. Inf. Syst.; 2012; 33, pp. 1-33. [DOI: https://dx.doi.org/10.1007/s10115-011-0463-8]
9. Luengo, J.; García-Gil, D.; Ramírez-Gallego, S.; García, S.; Herrera, F. Big Data Preprocessing; Springer: Cham, Switzerland, 2020.
10. Shen, C.; Zhang, K. Two-stage improved Grey Wolf optimization algorithm for feature selection on high-dimensional classification. Complex Intell. Syst.; 2021; 8, pp. 2769-2789. [DOI: https://dx.doi.org/10.1007/s40747-021-00452-4]
11. Fu, W.; Wang, K.; Tan, J.; Zhang, K. A composite framework coupling multiple feature selection, compound prediction models and novel hybrid swarm optimizer-based synchronization optimization strategy for multi-step ahead short-term wind speed forecasting. Energy Convers. Manag.; 2020; 205, 112461. [DOI: https://dx.doi.org/10.1016/j.enconman.2019.112461]
12. Di Mauro, M.; Galatro, G.; Fortino, G.; Liotta, A. Supervised feature selection techniques in network intrusion detection: A critical review. Eng. Appl. Artif. Intell.; 2021; 101, 104216. [DOI: https://dx.doi.org/10.1016/j.engappai.2021.104216]
13. Kashef, S.; Nezamabadi-pour, H.; Nikpour, B. Multilabel feature selection: A comprehensive review and guiding experiments. Wiley Interdiscip. Rev. Data Min. Knowl. Discov.; 2018; 8, e1240. [DOI: https://dx.doi.org/10.1002/widm.1240]
14. Zheng, Q.; Yang, M.; Tian, X.; Jiang, N.; Wang, D. A full stage data augmentation method in deep convolutional neural network for natural image classification. Discrete Dyn. Nat. Soc.; 2020; 2020, 4706576. [DOI: https://dx.doi.org/10.1155/2020/4706576]
15. Lee, C.Y.; Hung, C.H. Feature ranking and differential evolution for feature selection in brushless DC motor fault diagnosis. Symmetry; 2021; 13, 1291. [DOI: https://dx.doi.org/10.3390/sym13071291]
16. Li, J.; Gao, Y.; Wang, K.; Sun, Y. A dual opposition-based learning for differential evolution with protective mechanism for engineering optimization problems. Appl. Soft Comput.; 2021; 113, 107942. [DOI: https://dx.doi.org/10.1016/j.asoc.2021.107942]
17. Tsamardinos, I.; Charonyktakis, P.; Papoutsoglou, G.; Borboudakis, G.; Lakiotaki, K.; Zenklusen, J.C.; Juhl, H.; Chatzaki, E.; Lagani, V. Just Add Data: Automated predictive modeling for knowledge discovery and feature selection. NPJ Precis. Oncol.; 2022; 6, 38. [DOI: https://dx.doi.org/10.1038/s41698-022-00274-8]
18. Song, Y.; Wei, L.; Yang, Q.; Wu, J.; Xing, L.; Chen, Y. RL-GA: A reinforcement learning-based genetic algorithm for electromagnetic detection satellite scheduling problem. Swarm Evol. Comput.; 2023; 77, 101236. [DOI: https://dx.doi.org/10.1016/j.swevo.2023.101236]
19. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res.; 2003; 3, pp. 1157-1182.
20. Zhang, J.; Lin, Y.; Jiang, M.; Li, S.; Tang, Y.; Tan, K.C. Multi-label Feature Selection via Global Relevance and Redundancy Optimization. Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI); Yokohama, Japan, 7–15 January 2020; pp. 2512-2518.
21. Xue, B.; Zhang, M.; Browne, W.N. Particle swarm optimisation for feature selection in classification: Novel initialisation and updating mechanisms. Appl. Soft Comput.; 2014; 18, pp. 261-276.
22. Diao, R.; Shen, Q. Nature inspired feature selection meta-heuristics. Artif. Intell. Rev.; 2015; 44, pp. 311-340.
23. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell.; 2005; 27, pp. 1226-1238. [DOI: https://dx.doi.org/10.1109/TPAMI.2005.159]
24. Park, C.H.; Kim, S.B. Sequential random k-nearest neighbor feature selection for high-dimensional data. Expert Syst. Appl.; 2015; 42, pp. 2336-2342. [DOI: https://dx.doi.org/10.1016/j.eswa.2014.10.044]
25. Oh, I.S.; Lee, J.S.; Moon, B.R. Hybrid genetic algorithms for feature selection. IEEE Trans. Pattern Anal. Mach. Intell.; 2004; 26, pp. 1424-1437. [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/15521491]
26. Du, G.; Zhang, J.; Luo, Z.; Ma, F.; Ma, L.; Li, S. Joint imbalanced classification and feature selection for hospital readmissions. Knowl.-Based Syst.; 2020; 200, 106020. [DOI: https://dx.doi.org/10.1016/j.knosys.2020.106020]
27. Zhao, M.; Jha, A.; Liu, Q.; Millis, B.A.; Mahadevan-Jansen, A.; Lu, L.; Landman, B.A.; Tyska, M.J.; Huo, Y. Faster Mean-shift: GPU-accelerated clustering for cosine embedding-based cell segmentation and tracking. Med. Image Anal.; 2021; 71, 102048. [DOI: https://dx.doi.org/10.1016/j.media.2021.102048] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33872961]
28. Zhao, M.; Chang, C.H.; Xie, W.; Xie, Z.; Hu, J. Cloud shape classification system based on multi-channel cnn and improved fdm. IEEE Access; 2020; 8, pp. 44111-44124. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.2978090]
29. Zimbardo, G.; Malara, F.; Perri, S. Energetic particle superdiffusion in solar system plasmas: Which fractional transport equation?. Symmetry; 2021; 13, 2368. [DOI: https://dx.doi.org/10.3390/sym13122368]
30. Bi, Y.; Xue, B.; Mesejo, P.; Cagnoni, S.; Zhang, M. A Survey on Evolutionary Computation for Computer Vision and Image Analysis: Past, Present, and Future Trends. arXiv; 2022; arXiv: 2209.06399[DOI: https://dx.doi.org/10.1109/TEVC.2022.3220747]
31. Xu, J.; Sun, Y.; Qu, K.; Meng, X.; Hou, Q. Online group streaming feature selection using entropy-based uncertainty measures for fuzzy neighborhood rough sets. Complex Intell. Syst.; 2022; 8, pp. 5309-5328. [DOI: https://dx.doi.org/10.1007/s40747-022-00763-0]
32. Chen, L.Q.; Wang, C.; Song, S.L. Software defect prediction based on nested-stacking and heterogeneous feature selection. Complex Intell. Syst.; 2022; 8, pp. 3333-3348. [DOI: https://dx.doi.org/10.1007/s40747-022-00676-y]
33. Xu, J.; Yuan, M.; Ma, Y. Feature selection using self-information and entropy-based uncertainty measure for fuzzy neighborhood rough set. Complex Intell. Syst.; 2021; 8, pp. 287-305. [DOI: https://dx.doi.org/10.1007/s40747-021-00356-3]
34. Jain, R.; Joseph, T.; Saxena, A.; Gupta, D.; Khanna, A.; Sagar, K.; Ahlawat, A.K. Feature selection algorithm for usability engineering: A nature inspired approach. Complex Intell. Syst.; 2021; pp. 1-11. [DOI: https://dx.doi.org/10.1007/s40747-021-00384-z]
35. Jin, B.; Cruz, L.; Gonçalves, N. Deep facial diagnosis: Deep transfer learning from face recognition to facial diagnosis. IEEE Access; 2020; 8, pp. 123649-123661. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3005687]
36. Emary, E.; Zawbaa, H.M.; Hassanien, A.E. Binary grey wolf optimization approaches for feature selection. Neurocomputing; 2016; 172, pp. 371-381. [DOI: https://dx.doi.org/10.1016/j.neucom.2015.06.083]
37. Djemame, S.; Batouche, M.; Oulhadj, H.; Siarry, P. Solving reverse emergence with quantum PSO application to image processing. Soft Comput.; 2019; 23, pp. 6921-6935. [DOI: https://dx.doi.org/10.1007/s00500-018-3331-6]
38. Hosseini, S.; Zade, B.M.H. New hybrid method for attack detection using combination of evolutionary algorithms, SVM, and ANN. Comput. Netw.; 2020; 173, 107168. [DOI: https://dx.doi.org/10.1016/j.comnet.2020.107168]
39. Wu, H.; Gao, Y.; Wang, W.; Zhang, Z. A hybrid ant colony algorithm based on multiple strategies for the vehicle routing problem with time windows. Complex Intell. Syst.; 2021; pp. 1-18. [DOI: https://dx.doi.org/10.1007/s40747-021-00401-1]
40. Moghaddasi, S.S.; Faraji, N. A hybrid algorithm based on particle filter and genetic algorithm for target tracking. Expert Syst. Appl.; 2020; 147, 113188. [DOI: https://dx.doi.org/10.1016/j.eswa.2020.113188]
41. Hamdi, T.; Ali, J.B.; Di Costanzo, V.; Fnaiech, F.; Moreau, E.; Ginoux, J.M. Accurate prediction of continuous blood glucose based on support vector regression and differential evolution algorithm. Biocybern. Biomed. Eng.; 2018; 38, pp. 362-372. [DOI: https://dx.doi.org/10.1016/j.bbe.2018.02.005]
42. Euchi, J.; Masmoudi, M.; Siarry, P. Home health care routing and scheduling problems: A literature review. 4OR; 2022; 20, pp. 351-389. [DOI: https://dx.doi.org/10.1007/s10288-022-00516-2]
43. Harizan, S.; Kuila, P. Evolutionary algorithms for coverage and connectivity problems in wireless sensor networks: A study. Design Frameworks for Wireless Networks; Springer: Berlin/Heidelberg, Germany, 2020; pp. 257-280.
44. Mirjalili, S. Evolutionary algorithms and neural networks. Studies in Computational Intelligence; Springer: Berlin/Heidelberg, Germany, 2019; Volume 780.
45. Kamath, U.; Compton, J.; Islamaj-Doğan, R.; De Jong, K.A.; Shehu, A. An evolutionary algorithm approach for feature generation from sequence data and its application to DNA splice site prediction. IEEE/ACM Trans. Comput. Biol. Bioinform.; 2012; 9, pp. 1387-1398. [DOI: https://dx.doi.org/10.1109/TCBB.2012.53]
46. Abd-Alsabour, N. A review on evolutionary feature selection. Proceedings of the 2014 European Modelling Symposium; Pisa, Italy, 21–23 October 2014; pp. 20-26.
47. Jadhav, S.; He, H.; Jenkins, K. Information gain directed genetic algorithm wrapper feature selection for credit rating. Appl. Soft Comput.; 2018; 69, pp. 541-553. [DOI: https://dx.doi.org/10.1016/j.asoc.2018.04.033]
48. Ghamisi, P.; Benediktsson, J.A. Feature selection based on hybridization of genetic algorithm and particle swarm optimization. IEEE Geosci. Remote Sens. Lett.; 2014; 12, pp. 309-313. [DOI: https://dx.doi.org/10.1109/LGRS.2014.2337320]
49. Wang, X.; Yang, J.; Teng, X.; Xia, W.; Jensen, R. Feature selection based on rough sets and particle swarm optimization. Pattern Recognit. Lett.; 2007; 28, pp. 459-471. [DOI: https://dx.doi.org/10.1016/j.patrec.2006.09.003]
50. Braik, M.; Hammouri, A.; Atwan, J.; Al-Betar, M.A.; Awadallah, M.A. White Shark Optimizer: A novel bio-inspired meta-heuristic algorithm for global optimization problems. Knowl.-Based Syst.; 2022; 243, 108457. [DOI: https://dx.doi.org/10.1016/j.knosys.2022.108457]
51. Xue, B.; Zhang, M.; Browne, W.N.; Yao, X. A survey on evolutionary computation approaches to feature selection. IEEE Trans. Evol. Comput.; 2015; 20, pp. 606-626. [DOI: https://dx.doi.org/10.1109/TEVC.2015.2504420]
52. Maleki, N.; Zeinali, Y.; Niaki, S.T.A. A k-NN method for lung cancer prognosis with the use of a genetic algorithm for feature selection. Expert Syst. Appl.; 2021; 164, 113981. [DOI: https://dx.doi.org/10.1016/j.eswa.2020.113981]
53. Zhou, Y.; Zhang, W.; Kang, J.; Zhang, X.; Wang, X. A problem-specific non-dominated sorting genetic algorithm for supervised feature selection. Inf. Sci.; 2021; 547, pp. 841-859. [DOI: https://dx.doi.org/10.1016/j.ins.2020.08.083]
54. Xue, Y.; Zhu, H.; Liang, J.; Słowik, A. Adaptive crossover operator based multi-objective binary genetic algorithm for feature selection in classification. Knowl.-Based Syst.; 2021; 227, 107218. [DOI: https://dx.doi.org/10.1016/j.knosys.2021.107218]
55. Song, X.f.; Zhang, Y.; Gong, D.w.; Sun, X.y. Feature selection using bare-bones particle swarm optimization with mutual information. Pattern Recognit.; 2021; 112, 107804. [DOI: https://dx.doi.org/10.1016/j.patcog.2020.107804]
56. Song, X.F.; Zhang, Y.; Gong, D.W.; Gao, X.Z. A fast hybrid feature selection based on correlation-guided clustering and particle swarm optimization for high-dimensional data. IEEE Trans. Cybern.; 2021; 52, pp. 9573-9586. [DOI: https://dx.doi.org/10.1109/TCYB.2021.3061152]
57. Li, A.D.; Xue, B.; Zhang, M. Improved binary particle swarm optimization for feature selection with new initialization and search space reduction strategies. Appl. Soft Comput.; 2021; 106, 107302. [DOI: https://dx.doi.org/10.1016/j.asoc.2021.107302]
58. Jangir, P.; Jangir, N. A new non-dominated sorting grey wolf optimizer (NS-GWO) algorithm: Development and application to solve engineering designs and economic constrained emission dispatch problem with integration of wind power. Eng. Appl. Artif. Intell.; 2018; 72, pp. 449-467. [DOI: https://dx.doi.org/10.1016/j.engappai.2018.04.018]
59. Sathiyabhama, B.; Kumar, S.U.; Jayanthi, J.; Sathiya, T.; Ilavarasi, A.; Yuvarajan, V.; Gopikrishna, K. A novel feature selection framework based on grey wolf optimizer for mammogram image analysis. Neural Comput. Appl.; 2021; 33, pp. 14583-14602. [DOI: https://dx.doi.org/10.1007/s00521-021-06099-z]
60. Chen, H.; Ma, X.; Huang, S. A Feature Selection Method for Intrusion Detection Based on Parallel Sparrow Search Algorithm. Proceedings of the 2021 16th International Conference on Computer Science & Education (ICCSE); Lancaster, UK, 17–21 August 2021; pp. 685-690.
61. Da Silva, R.G.; Ribeiro, M.H.D.M.; Mariani, V.C.; dos Santos Coelho, L. Forecasting Brazilian and American COVID-19 cases based on artificial intelligence coupled with climatic exogenous variables. Chaos Solitons Fractals; 2020; 139, 110027. [DOI: https://dx.doi.org/10.1016/j.chaos.2020.110027] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32834591]
62. Dey, A.; Chattopadhyay, S.; Singh, P.K.; Ahmadian, A.; Ferrara, M.; Senu, N.; Sarkar, R. MRFGRO: A hybrid meta-heuristic feature selection method for screening COVID-19 using deep features. Sci. Rep.; 2021; 11, 24065. [DOI: https://dx.doi.org/10.1038/s41598-021-02731-z]
63. Shaban, W.M.; Rabie, A.H.; Saleh, A.I.; Abo-Elsoud, M. Accurate detection of COVID-19 patients based on distance biased Naïve Bayes (DBNB) classification strategy. Pattern Recognit.; 2021; 119, 108110. [DOI: https://dx.doi.org/10.1016/j.patcog.2021.108110]
64. Adam, S.P.; Alexandropoulos, S.A.N.; Pardalos, P.M.; Vrahatis, M.N. No free lunch theorem: A review. Approximation and Optimization; Springer: Berlin, Germany, 2019; pp. 57-82. [DOI: https://dx.doi.org/10.1007/978-3-030-12767-1_5]
65. Liu, T.; Yuan, Z.; Wu, L.; Badami, B. An optimal brain tumor detection by convolutional neural network and enhanced sparrow search algorithm. Proc. Inst. Mech. Eng. Part H J. Eng. Med.; 2021; 235, pp. 459-469. [DOI: https://dx.doi.org/10.1177/0954411920987964]
66. Zhu, Y.; Yousefi, N. Optimal parameter identification of PEMFC stacks using Adaptive Sparrow Search Algorithm. Int. J. Hydrogen Energy; 2021; 46, pp. 9541-9552. [DOI: https://dx.doi.org/10.1016/j.ijhydene.2020.12.107]
67. Zhang, C.; Ding, S. A stochastic configuration network based on chaotic sparrow search algorithm. Knowl.-Based Syst.; 2021; 220, 106924. [DOI: https://dx.doi.org/10.1016/j.knosys.2021.106924]
68. Tuerxun, W.; Chang, X.; Hongyu, G.; Zhijie, J.; Huajian, Z. Fault diagnosis of wind turbines based on a support vector machine optimized by the sparrow search algorithm. IEEE Access; 2021; 9, pp. 69307-69315. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3075547]
69. Gad, A.G.; Sallam, K.M.; Chakrabortty, R.K.; Ryan, M.J.; Abohany, A.A. An improved binary sparrow search algorithm for feature selection in data classification. Neural Comput. Appl.; 2022; 34, pp. 15705-15752. [DOI: https://dx.doi.org/10.1007/s00521-022-07203-7]
70. Xue, J.; Shen, B. A novel swarm intelligence optimization approach: Sparrow search algorithm. Syst. Sci. Control Eng.; 2020; 8, pp. 22-34. [DOI: https://dx.doi.org/10.1080/21642583.2019.1708830]
71. Wu, R.; Huang, H.; Wei, J.; Ma, C.; Zhu, Y.; Chen, Y.; Fan, Q. An improved sparrow search algorithm based on quantum computations and multi-strategy enhancement. Expert Syst. Appl.; 2023; 215, 119421. [DOI: https://dx.doi.org/10.1016/j.eswa.2022.119421]
72. Ma, J.; Hao, Z.; Sun, W. Enhancing sparrow search algorithm via multi-strategies for continuous optimization problems. Inf. Process. Manag.; 2022; 59, 102854. [DOI: https://dx.doi.org/10.1016/j.ipm.2021.102854]
73. Wang, P.; Zhang, Y.; Yang, H. Research on economic optimization of microgrid cluster based on chaos sparrow search algorithm. Comput. Intell. Neurosci.; 2021; 2021, 5556780. [DOI: https://dx.doi.org/10.1155/2021/5556780]
74. Zhang, N.; Zhao, Z.; Bao, X.; Qian, J.; Wu, B. Gravitational search algorithm based on improved Tent chaos. Control Decis.; 2020; 35, pp. 893-900.
75. Kuang, F.; Xu, W.; Jin, Z. Artificial bee colony algorithm based on self-adaptive Tent chaos search. Control Theory Appl.; 2014; 31, pp. 1502-1509.
76. Shan, L.; Qiang, H.; Li, J.; Wang, Z. Chaotic optimization algorithm based on Tent map. Control Decis.; 2005; 20, pp. 179-182.
77. Yang, X.S. Firefly algorithm, Levy flights and global optimization. Research and Development in Intelligent Systems XXVI; Springer: Berlin/Heidelberg, Germany, 2010; pp. 209-218.
78. Cao, W.; Tan, Y.; Huang, M.; Luo, Y. Adaptive bacterial foraging optimization based on roulette strategy. Proceedings of the International Conference on Swarm Intelligence; Barcelona, Spain, 26–28 October 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 299-311.
79. Altman, N.S. An introduction to kernel and nearest-neighbor nonparametric regression. Am. Stat.; 1992; 46, pp. 175-185.
80. Suganthan, P.N.; Hansen, N.; Liang, J.J.; Deb, K.; Chen, Y.P.; Auger, A.; Tiwari, S. Problem definitions and evaluation criteria for the CEC 2005 special session on real-parameter optimization. KanGAL Rep.; 2005; 2005005, 2005.
81. Tang, K.; Yáo, X.; Suganthan, P.N.; MacNish, C.; Chen, Y.P.; Chen, C.M.; Yang, Z. Benchmark Functions for the CEC’2008 Special Session and Competition on Large Scale Global Optimization; Nature Inspired Computation and Applications Laboratory, USTC: Beijing, China, 2007; Volume 24, pp. 1-18.
82. Mallipeddi, R.; Suganthan, P.N. Problem Definitions and Evaluation Criteria for the CEC 2010 Competition on Constrained Real-Parameter Optimization; Nanyang Technological University: Singapore, 2010; Volume 24.
83. Liang, J.J.; Qu, B.Y.; Suganthan, P.N. Problem Definitions and Evaluation Criteria for the CEC 2014 Special Session and Competition on Single Objective Real-Parameter Numerical Optimization; Technical Report Computational Intelligence Laboratory, Zhengzhou University: Zhengzhou, China, Nanyang Technological University: Singapore, 2013; Volume 635, 490.
84. Liang, J.; Qu, B.; Suganthan, P.; Chen, Q. Problem Definitions and Evaluation Criteria for the CEC 2015 Competition on Learning-Based Real-Parameter Single Objective Optimization; Technical Report 201411A Computational Intelligence Laboratory, Zhengzhou University: Zhengzhou, China, Nanyang Technological University: Singapore, 2014; Volume 29, pp. 625-640.
85. Wu, G.; Mallipeddi, R.; Suganthan, P.N. Problem Definitions and Evaluation Criteria for the CEC 2017 Competition on Constrained Real-Parameter Optimization; Technical Report National University of Defense Technology: Changsha, China, Kyungpook National University: Daegu, Republic of Korea, Nanyang Technological University: Singapore, 2017.
86. Mohamed, A.W.; Hadi, A.A.; Mohamed, A.K.; Awad, N.H. Evaluating the performance of adaptive GainingSharing knowledge based algorithm on CEC 2020 benchmark problems. Proceedings of the 2020 IEEE Congress on Evolutionary Computation (CEC); Glasgow, UK, 19–24 July 2020; pp. 1-8.
87. Yao, X.; Liu, Y.; Lin, G. Evolutionary programming made faster. IEEE Trans. Evol. Comput.; 1999; 3, pp. 82-102.
88. Karaboga, D.; Akay, B. A comparative study of artificial bee colony algorithm. Appl. Math. Comput.; 2009; 214, pp. 108-132. [DOI: https://dx.doi.org/10.1016/j.amc.2009.03.090]
89. Kennedy, J.; Eberhart, R. Particle swarm optimization. Proceedings of the ICNN’95-International Conference on Neural Networks; Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942-1948.
90. Cheng, R.; Jin, Y. A competitive swarm optimizer for large scale optimization. IEEE Trans. Cybern.; 2014; 45, pp. 191-204. [DOI: https://dx.doi.org/10.1109/TCYB.2014.2322602] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/24860047]
91. Liu, J.; Lampinen, J. A fuzzy adaptive differential evolution algorithm. Soft Comput.; 2005; 9, pp. 448-462. [DOI: https://dx.doi.org/10.1007/s00500-004-0363-x]
92. Zhu, G.Y.; Zhang, W.B. Optimal foraging algorithm for global optimization. Appl. Soft Comput.; 2017; 51, pp. 294-313. [DOI: https://dx.doi.org/10.1016/j.asoc.2016.11.047]
93. Viktorin, A.; Pluhacek, M.; Senkerik, R. Success-history based adaptive differential evolution algorithm with multi-chaotic framework for parent selection performance on CEC2014 benchmark set. Proceedings of the 2016 IEEE Congress on Evolutionary Computation (CEC); Vancouver, BC, Canada, 24–29 July 2016; pp. 4797-4803.
94. Li, J.; Gao, Y.; Zhang, H.; Yang, Q. Self-adaptive opposition-based differential evolution with subpopulation strategy for numerical and engineering optimization problems. Complex Intell. Syst.; 2022; 8, pp. 2051-2089. [DOI: https://dx.doi.org/10.1007/s40747-022-00734-5]
95. Asuncion, A.; Newman, D. UCI Machine Learning Repository; Irvine University of California: Irvine, CA, USA, 2007.
96. Holland, J.H. Genetic algorithms. Sci. Am.; 1992; 267, pp. 66-73. [DOI: https://dx.doi.org/10.1038/scientificamerican0792-66]
97. Mirjalili, S. Dragonfly algorithm: A new meta-heuristic optimization technique for solving single-objective, discrete, and multi-objective problems. Neural Comput. Appl.; 2016; 27, pp. 1053-1073. [DOI: https://dx.doi.org/10.1007/s00521-015-1920-1]
98. Mirjalili, S. The ant lion optimizer. Adv. Eng. Softw.; 2015; 83, pp. 80-98. [DOI: https://dx.doi.org/10.1016/j.advengsoft.2015.01.010]
99. Mirjalili, S. SCA: A sine cosine algorithm for solving optimization problems. Knowl.-Based Syst.; 2016; 96, pp. 120-133. [DOI: https://dx.doi.org/10.1016/j.knosys.2015.12.022]
100. Arora, S.; Anand, P. Binary butterfly optimization approaches for feature selection. Expert Syst. Appl.; 2019; 116, pp. 147-160. [DOI: https://dx.doi.org/10.1016/j.eswa.2018.08.051]
101. Shi, Y. Brain storm optimization algorithm. Proceedings of the International Conference in Swarm Intelligence; Chongqing, China, 12–15 June 2011; Springer: Berlin/Heidelberg, Germany, 2011; pp. 303-309.
102. Yuan, J.; Zhao, Z.; Liu, Y.; He, B.; Wang, L.; Xie, B.; Gao, Y. DMPPT control of photovoltaic microgrid based on improved sparrow search algorithm. IEEE Access; 2021; 9, pp. 16623-16629. [DOI: https://dx.doi.org/10.1109/ACCESS.2021.3052960]
103. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw.; 2014; 69, pp. 46-61. [DOI: https://dx.doi.org/10.1016/j.advengsoft.2013.12.007]
104. Wilcoxon, F. Individual comparisons by ranking methods. Breakthroughs in Statistics; Springer: Berlin/Heidelberg, Germany, 1992; pp. 196-202.
105. Derrac, J.; García, S.; Molina, D.; Herrera, F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol. Comput.; 2011; 1, pp. 3-18. [DOI: https://dx.doi.org/10.1016/j.swevo.2011.02.002]
106. Sayed, A.M.; Khattab, A.R.; AboulMagd, A.M.; Hassan, H.M.; Rateb, M.E.; Zaid, H.; Abdelmohsen, U.R. Nature as a treasure trove of potential anti-SARS-CoV drug leads: A structural/mechanistic rationale. RSC Adv.; 2020; 10, pp. 19790-19802. [DOI: https://dx.doi.org/10.1039/D0RA04199H]
107. Chen, X.; Tang, Y.; Mo, Y.; Li, S.; Lin, D.; Yang, Z.; Yang, Z.; Sun, H.; Qiu, J.; Liao, Y. et al. A diagnostic model for coronavirus disease 2019 (COVID-19) based on radiological semantic and clinical features: A multi-center study. Eur. Radiol.; 2020; 30, pp. 4893-4902. [DOI: https://dx.doi.org/10.1007/s00330-020-06829-2]
108. Iwendi, C.; Bashir, A.K.; Peshkar, A.; Sujatha, R.; Chatterjee, J.M.; Pasupuleti, S.; Mishra, R.; Pillai, S.; Jo, O. COVID-19 patient health prediction using boosted random forest algorithm. Front. Public Health; 2020; 8, 357. [DOI: https://dx.doi.org/10.3389/fpubh.2020.00357]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
The “Curse of Dimensionality” induced by the rapid development of information science might have a negative impact when dealing with big datasets, and it also makes the problems of symmetry and asymmetry increasingly prominent. Feature selection (FS) can eliminate irrelevant information in big data and improve accuracy. As a recently proposed algorithm, the Sparrow Search Algorithm (SSA) shows its advantages in the FS tasks because of its superior performance. However, SSA is more subject to the population’s poor diversity and falls into a local optimum. Regarding this issue, we propose a variant of the SSA called the Tent Lévy Flying Sparrow Search Algorithm (TFSSA) to select the best subset of features in the wrapper-based method for classification purposes. After the performance results are evaluated on the CEC2020 test suite, TFSSA is used to select the best feature combination to maximize classification accuracy and simultaneously minimize the number of selected features. To evaluate the proposed TFSSA, we have conducted experiments on twenty-one datasets from the UCI repository to compare with nine algorithms in the literature. Nine metrics are used to evaluate and compare these algorithms’ performance properly. Furthermore, the method is also used on the coronavirus disease (COVID-19) dataset, and its classification accuracy and the average number of feature selections are 93.47% and 2.1, respectively, reaching the best. The experimental results and comparison in all datasets demonstrate the effectiveness of our new algorithm, TFSSA, compared with other wrapper-based algorithms.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details



1 School of Computer Science and Engineering, North Minzu University, Yinchuan 750021, China
2 Ningxia Key Laboratory of Intelligent Information and Big Data Processing, Yinchuan 750021, China
3 College of Systems Engineering, National University of Defense Technology, Changsha 410073, China